mradermacher
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -192,6 +192,10 @@ tags:
|
|
192 |
<!-- ### tags: nicoboss -->
|
193 |
weighted/imatrix quants of https://huggingface.co/meta-llama/Meta-Llama-3.1-405B-Instruct
|
194 |
|
|
|
|
|
|
|
|
|
195 |
<!-- provided-files -->
|
196 |
static quants are available at https://huggingface.co/mradermacher/Meta-Llama-3.1-405B-Instruct-GGUF
|
197 |
## Usage
|
@@ -211,7 +215,7 @@ more details, including on how to concatenate multi-part files.
|
|
211 |
Here is a handy graph by ikawrakow comparing some lower-quality quant
|
212 |
types (lower is better):
|
213 |
|
214 |
-
![image.png](
|
215 |
|
216 |
And here are Artefact2's thoughts on the matter:
|
217 |
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
|
|
|
192 |
<!-- ### tags: nicoboss -->
|
193 |
weighted/imatrix quants of https://huggingface.co/meta-llama/Meta-Llama-3.1-405B-Instruct
|
194 |
|
195 |
+
These imatrix quants have recently been requantized from a higher quality imatrix calculated
|
196 |
+
from the source model instead of the Q8_0, in what was probably the largest distributed imatrix
|
197 |
+
computation to date (and also one of the first).
|
198 |
+
|
199 |
<!-- provided-files -->
|
200 |
static quants are available at https://huggingface.co/mradermacher/Meta-Llama-3.1-405B-Instruct-GGUF
|
201 |
## Usage
|
|
|
215 |
Here is a handy graph by ikawrakow comparing some lower-quality quant
|
216 |
types (lower is better):
|
217 |
|
218 |
+
![image.png]()
|
219 |
|
220 |
And here are Artefact2's thoughts on the matter:
|
221 |
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
|