xangxiong commited on
Commit
d8a5509
1 Parent(s): e4ac64c

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "gate_proj",
 
24
  "down_proj",
25
  "v_proj",
26
- "up_proj",
27
  "k_proj",
28
- "o_proj",
29
- "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "o_proj",
25
  "down_proj",
26
  "v_proj",
 
27
  "k_proj",
28
+ "gate_proj",
29
+ "up_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3a57c7ed0d273e7cef406e8359b5592a997e7aa582c76e1fc8ba3ded400e789
3
- size 4785795512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f30dcd194322082e80bc03606bb27048b275c49cce7807d6e012f18e863525ee
3
+ size 2684416208
special_tokens_map.json CHANGED
@@ -1,21 +1,17 @@
1
  {
2
- "additional_special_tokens": [
3
- {
4
- "content": "<|im_start|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "<|im_end|>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- }
17
- ],
18
- "bos_token": "<|im_start|>",
19
- "eos_token": "<|im_end|>",
20
- "pad_token": "<|im_end|>"
21
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|end_of_text|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|end_of_text|>"
 
 
 
 
17
  }
tokenizer.json CHANGED
@@ -2306,24 +2306,6 @@
2306
  "rstrip": false,
2307
  "normalized": false,
2308
  "special": true
2309
- },
2310
- {
2311
- "id": 128256,
2312
- "content": "<|im_start|>",
2313
- "single_word": false,
2314
- "lstrip": false,
2315
- "rstrip": false,
2316
- "normalized": false,
2317
- "special": true
2318
- },
2319
- {
2320
- "id": 128257,
2321
- "content": "<|im_end|>",
2322
- "single_word": false,
2323
- "lstrip": false,
2324
- "rstrip": false,
2325
- "normalized": false,
2326
- "special": true
2327
  }
2328
  ],
2329
  "normalizer": null,
 
2306
  "rstrip": false,
2307
  "normalized": false,
2308
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2309
  }
2310
  ],
2311
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -2047,37 +2047,16 @@
2047
  "rstrip": false,
2048
  "single_word": false,
2049
  "special": true
2050
- },
2051
- "128256": {
2052
- "content": "<|im_start|>",
2053
- "lstrip": false,
2054
- "normalized": false,
2055
- "rstrip": false,
2056
- "single_word": false,
2057
- "special": true
2058
- },
2059
- "128257": {
2060
- "content": "<|im_end|>",
2061
- "lstrip": false,
2062
- "normalized": false,
2063
- "rstrip": false,
2064
- "single_word": false,
2065
- "special": true
2066
  }
2067
  },
2068
- "additional_special_tokens": [
2069
- "<|im_start|>",
2070
- "<|im_end|>"
2071
- ],
2072
- "bos_token": "<|im_start|>",
2073
- "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2074
  "clean_up_tokenization_spaces": true,
2075
- "eos_token": "<|im_end|>",
2076
  "model_input_names": [
2077
  "input_ids",
2078
  "attention_mask"
2079
  ],
2080
  "model_max_length": 1000000000000000019884624838656,
2081
- "pad_token": "<|im_end|>",
2082
  "tokenizer_class": "PreTrainedTokenizerFast"
2083
  }
 
2047
  "rstrip": false,
2048
  "single_word": false,
2049
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2050
  }
2051
  },
2052
+ "bos_token": "<|begin_of_text|>",
 
 
 
 
 
2053
  "clean_up_tokenization_spaces": true,
2054
+ "eos_token": "<|end_of_text|>",
2055
  "model_input_names": [
2056
  "input_ids",
2057
  "attention_mask"
2058
  ],
2059
  "model_max_length": 1000000000000000019884624838656,
2060
+ "pad_token": "<|end_of_text|>",
2061
  "tokenizer_class": "PreTrainedTokenizerFast"
2062
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3491da34df2bae7968044a345e1f14cff56501edfee6bb6f70463813d32f5eaf
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a21c2d1cea1fc4c4dc733fcdf93ad6ae38c0de23abab54da166994e3f1ca31d
3
  size 5432