NanoByte's picture
Update README.md
046a79c verified
|
raw
history blame contribute delete
No virus
814 Bytes
metadata
base_model: wolfram/miquliz-120b
inference: false
model_creator: Wolfram Ravenwolf
model_name: miquliz-120b

miquliz-120b - Q4 GGUF

Description

This repo contains Q4_K_S and Q4_K_M GGUF format model files for Wolfram Ravenwolf's miquliz-120b.

Prompt template: Mistral

[INST] {prompt} [/INST]

Provided files

Name Quant method Bits Size
miquliz-120b.Q4_K_S.gguf Q4_K_S 4 66.81 GB
miquliz-120b.Q4_K_M.gguf Q4_K_M 4 70.64 GB

Note: HF does not support uploading files larger than 50GB. Therefore the files are uploaded as split files.