royam0820 commited on
Commit
3ce4d0a
1 Parent(s): fc8b1f2

config.json

Browse files
Files changed (1) hide show
  1. config.json +25 -0
config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _name_or_path: "TheBloke/Llama-2-7B-fp16",
2
+ architectures: [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ bos_token_id: 1,
6
+ eos_token_id: 2,
7
+ hidden_act: "silu",
8
+ hidden_size: 4096,
9
+ initializer_range: 0.02,
10
+ intermediate_size: 11008,
11
+ max_position_embeddings: 2048,
12
+ model_type: "llama",
13
+ num_attention_heads: 32,
14
+ num_hidden_layers: 32,
15
+ num_key_value_heads: 32,
16
+ pad_token_id: 0,
17
+ pretraining_tp: 1,
18
+ rms_norm_eps: 0.00001,
19
+ rope_scaling: null,
20
+ tie_word_embeddings: false,
21
+ torch_dtype: "bfloat16",
22
+ transformers_version: "4.31.0",
23
+ use_cache: true,
24
+ vocab_size: 32000
25
+ }