Nickinfinity commited on
Commit
2be06e2
1 Parent(s): e346dcd

Update model

Browse files

Signed-off-by: Nickinfinity <nikhilg.me@gmail.com>

Files changed (1) hide show
  1. mlc-chat-config.json +7 -4
mlc-chat-config.json CHANGED
@@ -13,7 +13,8 @@
13
  "prefill_chunk_size": 768,
14
  "num_key_value_heads": 4,
15
  "head_dim": 64,
16
- "tensor_parallel_shards": 1
 
17
  },
18
  "vocab_size": 160984,
19
  "context_window_size": 768,
@@ -22,12 +23,14 @@
22
  "attention_sink_size": -1,
23
  "tensor_parallel_shards": 1,
24
  "mean_gen_len": 128,
25
- "max_gen_len": 49,
26
  "shift_fill_factor": 0.3,
27
  "temperature": 0.7,
 
 
28
  "repetition_penalty": 1.0,
29
  "top_p": 0.95,
30
- "conv_template": "llama_default",
31
  "pad_token_id": 0,
32
  "bos_token_id": 1,
33
  "eos_token_id": 2,
@@ -35,4 +38,4 @@
35
  "tokenizer.json"
36
  ],
37
  "version": "0.1.0"
38
- }
 
13
  "prefill_chunk_size": 768,
14
  "num_key_value_heads": 4,
15
  "head_dim": 64,
16
+ "tensor_parallel_shards": 1,
17
+ "max_batch_size": 80
18
  },
19
  "vocab_size": 160984,
20
  "context_window_size": 768,
 
23
  "attention_sink_size": -1,
24
  "tensor_parallel_shards": 1,
25
  "mean_gen_len": 128,
26
+ "max_gen_len": 512,
27
  "shift_fill_factor": 0.3,
28
  "temperature": 0.7,
29
+ "presence_penalty": 0.0,
30
+ "frequency_penalty": 0.0,
31
  "repetition_penalty": 1.0,
32
  "top_p": 0.95,
33
+ "conv_template": "llama-2",
34
  "pad_token_id": 0,
35
  "bos_token_id": 1,
36
  "eos_token_id": 2,
 
38
  "tokenizer.json"
39
  ],
40
  "version": "0.1.0"
41
+ }