teknium commited on
Commit
d8d5aaf
1 Parent(s): aaac35f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -9
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: NousResearch/Meta-Llama-3-8B
3
  tags:
4
  - Llama-3
5
  - instruct
@@ -14,7 +14,7 @@ tags:
14
  - json mode
15
  - axolotl
16
  model-index:
17
- - name: Hermes-2-Pro-Llama-3-8B
18
  results: []
19
  language:
20
  - en
@@ -33,7 +33,7 @@ widget:
33
  Buu to destroy the world.
34
  ---
35
 
36
- # Hermes 2 Pro - Llama-3 8B
37
 
38
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/ggO2sBDJ8Bhc6w-zwTx5j.png)
39
 
@@ -262,9 +262,9 @@ import torch
262
  from transformers import AutoTokenizer, AutoModelForCausalLM, LlamaForCausalLM
263
  import bitsandbytes, flash_attn
264
 
265
- tokenizer = AutoTokenizer.from_pretrained('NousResearch/Hermes-2-Pro-Llama-3-8B', trust_remote_code=True)
266
  model = LlamaForCausalLM.from_pretrained(
267
- "NousResearch/Hermes-2-Pro-Llama-3-8B",
268
  torch_dtype=torch.float16,
269
  device_map="auto",
270
  load_in_8bit=False,
@@ -306,14 +306,14 @@ In LM-Studio, simply select the ChatML Prefix on the settings side pane:
306
 
307
  ## Quantized Versions:
308
 
309
- GGUF Versions Available Here: https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF
310
 
311
  # How to cite:
312
 
313
  ```bibtext
314
- @misc{Hermes-2-Pro-Llama-3-8B,
315
- url={[https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B]https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B)},
316
- title={Hermes-2-Pro-Llama-3-8B},
317
  author={"Teknium", "interstellarninja", "theemozilla", "karan4d", "huemin_art"}
318
  }
319
  ```
 
1
  ---
2
+ base_model: NousResearch/Meta-Llama-3-70B
3
  tags:
4
  - Llama-3
5
  - instruct
 
14
  - json mode
15
  - axolotl
16
  model-index:
17
+ - name: Hermes-2-Pro-Llama-3-70B
18
  results: []
19
  language:
20
  - en
 
33
  Buu to destroy the world.
34
  ---
35
 
36
+ # Hermes 2 Pro - Llama-3 70B
37
 
38
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/ggO2sBDJ8Bhc6w-zwTx5j.png)
39
 
 
262
  from transformers import AutoTokenizer, AutoModelForCausalLM, LlamaForCausalLM
263
  import bitsandbytes, flash_attn
264
 
265
+ tokenizer = AutoTokenizer.from_pretrained('NousResearch/Hermes-2-Pro-Llama-3-70B', trust_remote_code=True)
266
  model = LlamaForCausalLM.from_pretrained(
267
+ "NousResearch/Hermes-2-Pro-Llama-3-70B",
268
  torch_dtype=torch.float16,
269
  device_map="auto",
270
  load_in_8bit=False,
 
306
 
307
  ## Quantized Versions:
308
 
309
+ GGUF Versions *will soon be* Available Here: https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-70B-GGUF
310
 
311
  # How to cite:
312
 
313
  ```bibtext
314
+ @misc{Hermes-2-Pro-Llama-3-70B,
315
+ url={[https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-70B]https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-70B)},
316
+ title={Hermes-2-Pro-Llama-3-70B},
317
  author={"Teknium", "interstellarninja", "theemozilla", "karan4d", "huemin_art"}
318
  }
319
  ```