habib-z commited on
Commit
c2dd8d6
1 Parent(s): 651fefc

Upload LlamaForCausalLM

Browse files
README.md CHANGED
@@ -1,6 +1,8 @@
1
  ---
2
  library_name: transformers
3
- tags: []
 
 
4
  ---
5
 
6
  # Model Card for Model ID
 
1
  ---
2
  library_name: transformers
3
+ tags:
4
+ - trl
5
+ - sft
6
  ---
7
 
8
  # Model Card for Model ID
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "NousResearch/Llama-2-7b-chat-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -37,8 +37,8 @@
37
  "rope_scaling": null,
38
  "rope_theta": 10000.0,
39
  "tie_word_embeddings": false,
40
- "torch_dtype": "float16",
41
- "transformers_version": "4.40.1",
42
  "use_cache": false,
43
  "vocab_size": 41510
44
  }
 
1
  {
2
+ "_name_or_path": "habib-z/llama-2-7b-chat-fa",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
37
  "rope_scaling": null,
38
  "rope_theta": 10000.0,
39
  "tie_word_embeddings": false,
40
+ "torch_dtype": "float32",
41
+ "transformers_version": "4.40.2",
42
  "use_cache": false,
43
  "vocab_size": 41510
44
  }
generation_config.json CHANGED
@@ -5,5 +5,5 @@
5
  "pad_token_id": 32000,
6
  "temperature": null,
7
  "top_p": null,
8
- "transformers_version": "4.40.1"
9
  }
 
5
  "pad_token_id": 32000,
6
  "temperature": null,
7
  "top_p": null,
8
+ "transformers_version": "4.40.2"
9
  }
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22b12a63e2baf22e3873490305f0ffaa2041d02c9e82cb44567a66b33bce6efc
3
+ size 4340863048
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efac40482a8dcd7df7dc9c95c7aec6d563be1264d5bf319fb20df12309864cff
3
+ size 680099968
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff