morriszms commited on
Commit
ea1a819
1 Parent(s): dbc7783

Upload folder using huggingface_hub

Browse files
Meta-Llama-3-8B-Q2_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c4a35e4297991bacde5011750974ee6adff1e74f79a3734cbc745c5530f00c4
3
- size 3179131200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f296d7867ecba4395032660e9b3777ae1a6ebf8c388f83e09b5528d59e22d5f2
3
+ size 3179131104
Meta-Llama-3-8B-Q3_K_L.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73b69ba64214a80f8aa6b9de4d44c97e712a303965cc67f6e1e6b7f22cab3860
3
- size 4321956160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af5c7a9d23c4f9690ff09f45a1b8d66d85dec1c2b7bf1218c683edfc76d456fa
3
+ size 4321956064
Meta-Llama-3-8B-Q3_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8e952df35646dd6185da38d36b6f8ec1093fea883870f3e99c4d0737bdd7fb9b
3
- size 4018917696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3db3b17fe5db4b761d4615ad219829f07cab1cf94962ed791ca21ffa5efa312c
3
+ size 4018917600
Meta-Llama-3-8B-Q3_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89e7e1c199f0112b44336c1f69472895d322c20f71f3f8146c18fd3902d32bc9
3
- size 3664499008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a95bcdc14ef3c68298c02e08c05f6e35ce00921a9a735d4a8dd13940f9647b8
3
+ size 3664498912
Meta-Llama-3-8B-Q4_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8155d4fc0b043a81fe8fe4ba7744f10543eb970f8f5351cc14e5f0f37ec8c29
3
- size 4661211456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dfa2743042a82d00bf09dc7ab800531896a0724352009534cd55e07a6653ce7
3
+ size 4661211360
Meta-Llama-3-8B-Q4_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87d402c4a018b27e70b80224f7490f40427f95748f4e1c91bd930a3665075956
3
- size 4920734016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:619b435df6c376f56a7b7527dd0290e153b9d0a891a928b231c074d71ad6bdd8
3
+ size 4920733920
Meta-Llama-3-8B-Q4_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6cdc2c8163d2bc4bfb4c88653f80180afa8e62aba493497ec33a17f8a607283
3
- size 4692668736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c785f18759c63f83e016f7474cc99dd27d29b0a78f45f08568148788884c710e
3
+ size 4692668640
Meta-Llama-3-8B-Q5_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a97876a23986ea319385269721c3d74f71bd7275cfa0e9763883c04e874d3df
3
- size 5599293760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6376e421da3b6a1836510810e2b198a126900af041e258fdb0ffd07cb600c6d
3
+ size 5599293664
Meta-Llama-3-8B-Q5_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8b2df4fc8ef0bb5a84603222d5fb9181bae75664e5f1cb1f557fe8f37a57f87
3
- size 5732987200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:815e00bc7a7f81ef5f879b0dea3c0dffbedae1342fd45d109bee8df4a39e080b
3
+ size 5732987104
Meta-Llama-3-8B-Q5_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:faaabf46f9c17d02cf6b9a56223351eb4c91ef53ecc542c35a52ac4ec0beafc8
3
- size 5599293760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e7c94e77df8ef4bfc3a4e17fb7d6d78e16551019d87920eb70606462f5c4fa1
3
+ size 5599293664
Meta-Llama-3-8B-Q6_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:05eb88328365b330bf6a33874913e228c6e5958e0fd7ee0cd9891c6e3023fe7b
3
- size 6596006208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6782f0b8e2d508010ee13c82d9dec9a371d148ee9f883d66ae2b12760d71ff3f
3
+ size 6596006112
Meta-Llama-3-8B-Q8_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df76c186cc88d8d05741b9d2161049671c88bab27fbbdcdcc86ab2562fb87068
3
- size 8540770624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9292538bf28a613dbec93b6db0628fbfebe5a21c3991c3851a02f2d988780c8
3
+ size 8540770528
README.md CHANGED
@@ -10,9 +10,8 @@ tags:
10
  - llama-3
11
  - TensorBlock
12
  - GGUF
13
- license: other
14
- license_name: llama3
15
- license_link: LICENSE
16
  extra_gated_prompt: "### META LLAMA 3 COMMUNITY LICENSE AGREEMENT\nMeta Llama 3 Version\
17
  \ Release Date: April 18, 2024\n\"Agreement\" means the terms and conditions for\
18
  \ use, reproduction, distribution and modification of the Llama Materials set forth\
@@ -177,7 +176,7 @@ extra_gated_fields:
177
  extra_gated_description: The information you provide will be collected, stored, processed
178
  and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
179
  extra_gated_button_content: Submit
180
- base_model: NousResearch/Meta-Llama-3-8B
181
  ---
182
 
183
  <div style="width: auto; margin-left: auto; margin-right: auto">
@@ -191,13 +190,12 @@ base_model: NousResearch/Meta-Llama-3-8B
191
  </div>
192
  </div>
193
 
194
- ## NousResearch/Meta-Llama-3-8B - GGUF
195
 
196
- This repo contains GGUF format model files for [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B).
197
 
198
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
199
 
200
-
201
  <div style="text-align: left; margin: 20px 0;">
202
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
203
  Run them on the TensorBlock client using your local machine ↗
@@ -206,7 +204,6 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
206
 
207
  ## Prompt template
208
 
209
-
210
  ```
211
 
212
  ```
@@ -215,18 +212,18 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
215
 
216
  | Filename | Quant type | File Size | Description |
217
  | -------- | ---------- | --------- | ----------- |
218
- | [Meta-Llama-3-8B-Q2_K.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q2_K.gguf) | Q2_K | 2.961 GB | smallest, significant quality loss - not recommended for most purposes |
219
- | [Meta-Llama-3-8B-Q3_K_S.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q3_K_S.gguf) | Q3_K_S | 3.413 GB | very small, high quality loss |
220
- | [Meta-Llama-3-8B-Q3_K_M.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q3_K_M.gguf) | Q3_K_M | 3.743 GB | very small, high quality loss |
221
- | [Meta-Llama-3-8B-Q3_K_L.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q3_K_L.gguf) | Q3_K_L | 4.025 GB | small, substantial quality loss |
222
- | [Meta-Llama-3-8B-Q4_0.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q4_0.gguf) | Q4_0 | 4.341 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
223
- | [Meta-Llama-3-8B-Q4_K_S.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q4_K_S.gguf) | Q4_K_S | 4.370 GB | small, greater quality loss |
224
- | [Meta-Llama-3-8B-Q4_K_M.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q4_K_M.gguf) | Q4_K_M | 4.583 GB | medium, balanced quality - recommended |
225
- | [Meta-Llama-3-8B-Q5_0.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q5_0.gguf) | Q5_0 | 5.215 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
226
- | [Meta-Llama-3-8B-Q5_K_S.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q5_K_S.gguf) | Q5_K_S | 5.215 GB | large, low quality loss - recommended |
227
- | [Meta-Llama-3-8B-Q5_K_M.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q5_K_M.gguf) | Q5_K_M | 5.339 GB | large, very low quality loss - recommended |
228
- | [Meta-Llama-3-8B-Q6_K.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q6_K.gguf) | Q6_K | 6.143 GB | very large, extremely low quality loss |
229
- | [Meta-Llama-3-8B-Q8_0.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q8_0.gguf) | Q8_0 | 7.954 GB | very large, extremely low quality loss - not recommended |
230
 
231
 
232
  ## Downloading instruction
 
10
  - llama-3
11
  - TensorBlock
12
  - GGUF
13
+ license: llama3
14
+ new_version: meta-llama/Llama-3.1-8B
 
15
  extra_gated_prompt: "### META LLAMA 3 COMMUNITY LICENSE AGREEMENT\nMeta Llama 3 Version\
16
  \ Release Date: April 18, 2024\n\"Agreement\" means the terms and conditions for\
17
  \ use, reproduction, distribution and modification of the Llama Materials set forth\
 
176
  extra_gated_description: The information you provide will be collected, stored, processed
177
  and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
178
  extra_gated_button_content: Submit
179
+ base_model: meta-llama/Meta-Llama-3-8B
180
  ---
181
 
182
  <div style="width: auto; margin-left: auto; margin-right: auto">
 
190
  </div>
191
  </div>
192
 
193
+ ## meta-llama/Meta-Llama-3-8B - GGUF
194
 
195
+ This repo contains GGUF format model files for [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B).
196
 
197
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
198
 
 
199
  <div style="text-align: left; margin: 20px 0;">
200
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
201
  Run them on the TensorBlock client using your local machine ↗
 
204
 
205
  ## Prompt template
206
 
 
207
  ```
208
 
209
  ```
 
212
 
213
  | Filename | Quant type | File Size | Description |
214
  | -------- | ---------- | --------- | ----------- |
215
+ | [Meta-Llama-3-8B-Q2_K.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q2_K.gguf) | Q2_K | 3.179 GB | smallest, significant quality loss - not recommended for most purposes |
216
+ | [Meta-Llama-3-8B-Q3_K_S.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q3_K_S.gguf) | Q3_K_S | 3.664 GB | very small, high quality loss |
217
+ | [Meta-Llama-3-8B-Q3_K_M.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q3_K_M.gguf) | Q3_K_M | 4.019 GB | very small, high quality loss |
218
+ | [Meta-Llama-3-8B-Q3_K_L.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q3_K_L.gguf) | Q3_K_L | 4.322 GB | small, substantial quality loss |
219
+ | [Meta-Llama-3-8B-Q4_0.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q4_0.gguf) | Q4_0 | 4.661 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
220
+ | [Meta-Llama-3-8B-Q4_K_S.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q4_K_S.gguf) | Q4_K_S | 4.693 GB | small, greater quality loss |
221
+ | [Meta-Llama-3-8B-Q4_K_M.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q4_K_M.gguf) | Q4_K_M | 4.921 GB | medium, balanced quality - recommended |
222
+ | [Meta-Llama-3-8B-Q5_0.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q5_0.gguf) | Q5_0 | 5.599 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
223
+ | [Meta-Llama-3-8B-Q5_K_S.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q5_K_S.gguf) | Q5_K_S | 5.599 GB | large, low quality loss - recommended |
224
+ | [Meta-Llama-3-8B-Q5_K_M.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q5_K_M.gguf) | Q5_K_M | 5.733 GB | large, very low quality loss - recommended |
225
+ | [Meta-Llama-3-8B-Q6_K.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q6_K.gguf) | Q6_K | 6.596 GB | very large, extremely low quality loss |
226
+ | [Meta-Llama-3-8B-Q8_0.gguf](https://huggingface.co/tensorblock/Meta-Llama-3-8B-GGUF/blob/main/Meta-Llama-3-8B-Q8_0.gguf) | Q8_0 | 8.541 GB | very large, extremely low quality loss - not recommended |
227
 
228
 
229
  ## Downloading instruction