File size: 458 Bytes
a633c81
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
---
language:
- en
license: apache-2.0
---

These are GGUF quantized versions of [perlthoughts/Mistral-11B-Instruct-v0.2](https://huggingface.co/perlthoughts/Mistral-11B-Instruct-v0.2).

The importance matrix was trained for 100K tokens (200 batches of 512 tokens) using `wiki.train.raw`.

Some model files above 50GB are split into smaller files. To concatenate them, use the `cat` command (on Windows, use PowerShell): `cat foo-Q6_K.gguf.* > foo-Q6_K.gguf`