Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
---
|
2 |
-
base_model: NousResearch/Meta-Llama-3-
|
3 |
tags:
|
4 |
- Llama-3
|
5 |
- instruct
|
@@ -14,7 +14,7 @@ tags:
|
|
14 |
- json mode
|
15 |
- axolotl
|
16 |
model-index:
|
17 |
-
- name: Hermes-2-Pro-Llama-3-
|
18 |
results: []
|
19 |
language:
|
20 |
- en
|
@@ -33,7 +33,7 @@ widget:
|
|
33 |
Buu to destroy the world.
|
34 |
---
|
35 |
|
36 |
-
# Hermes 2 Pro - Llama-3
|
37 |
|
38 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/ggO2sBDJ8Bhc6w-zwTx5j.png)
|
39 |
|
@@ -262,9 +262,9 @@ import torch
|
|
262 |
from transformers import AutoTokenizer, AutoModelForCausalLM, LlamaForCausalLM
|
263 |
import bitsandbytes, flash_attn
|
264 |
|
265 |
-
tokenizer = AutoTokenizer.from_pretrained('NousResearch/Hermes-2-Pro-Llama-3-
|
266 |
model = LlamaForCausalLM.from_pretrained(
|
267 |
-
"NousResearch/Hermes-2-Pro-Llama-3-
|
268 |
torch_dtype=torch.float16,
|
269 |
device_map="auto",
|
270 |
load_in_8bit=False,
|
@@ -306,14 +306,14 @@ In LM-Studio, simply select the ChatML Prefix on the settings side pane:
|
|
306 |
|
307 |
## Quantized Versions:
|
308 |
|
309 |
-
GGUF Versions Available Here: https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-
|
310 |
|
311 |
# How to cite:
|
312 |
|
313 |
```bibtext
|
314 |
-
@misc{Hermes-2-Pro-Llama-3-
|
315 |
-
url={[https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-
|
316 |
-
title={Hermes-2-Pro-Llama-3-
|
317 |
author={"Teknium", "interstellarninja", "theemozilla", "karan4d", "huemin_art"}
|
318 |
}
|
319 |
```
|
|
|
1 |
---
|
2 |
+
base_model: NousResearch/Meta-Llama-3-70B
|
3 |
tags:
|
4 |
- Llama-3
|
5 |
- instruct
|
|
|
14 |
- json mode
|
15 |
- axolotl
|
16 |
model-index:
|
17 |
+
- name: Hermes-2-Pro-Llama-3-70B
|
18 |
results: []
|
19 |
language:
|
20 |
- en
|
|
|
33 |
Buu to destroy the world.
|
34 |
---
|
35 |
|
36 |
+
# Hermes 2 Pro - Llama-3 70B
|
37 |
|
38 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/ggO2sBDJ8Bhc6w-zwTx5j.png)
|
39 |
|
|
|
262 |
from transformers import AutoTokenizer, AutoModelForCausalLM, LlamaForCausalLM
|
263 |
import bitsandbytes, flash_attn
|
264 |
|
265 |
+
tokenizer = AutoTokenizer.from_pretrained('NousResearch/Hermes-2-Pro-Llama-3-70B', trust_remote_code=True)
|
266 |
model = LlamaForCausalLM.from_pretrained(
|
267 |
+
"NousResearch/Hermes-2-Pro-Llama-3-70B",
|
268 |
torch_dtype=torch.float16,
|
269 |
device_map="auto",
|
270 |
load_in_8bit=False,
|
|
|
306 |
|
307 |
## Quantized Versions:
|
308 |
|
309 |
+
GGUF Versions *will soon be* Available Here: https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-70B-GGUF
|
310 |
|
311 |
# How to cite:
|
312 |
|
313 |
```bibtext
|
314 |
+
@misc{Hermes-2-Pro-Llama-3-70B,
|
315 |
+
url={[https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-70B]https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-70B)},
|
316 |
+
title={Hermes-2-Pro-Llama-3-70B},
|
317 |
author={"Teknium", "interstellarninja", "theemozilla", "karan4d", "huemin_art"}
|
318 |
}
|
319 |
```
|