benchang1110 commited on
Commit
db2be62
1 Parent(s): 6530cae

Upload tokenizer

Browse files
special_tokens_map.json CHANGED
@@ -18,7 +18,7 @@
18
  "single_word": false
19
  },
20
  "pad_token": {
21
- "content": "<pad>",
22
  "lstrip": false,
23
  "normalized": false,
24
  "rstrip": false,
 
18
  "single_word": false
19
  },
20
  "pad_token": {
21
+ "content": "<|im_end|>",
22
  "lstrip": false,
23
  "normalized": false,
24
  "rstrip": false,
tokenizer.json CHANGED
@@ -155,15 +155,6 @@
155
  "rstrip": false,
156
  "normalized": false,
157
  "special": true
158
- },
159
- {
160
- "id": 49152,
161
- "content": "<pad>",
162
- "single_word": false,
163
- "lstrip": false,
164
- "rstrip": false,
165
- "normalized": false,
166
- "special": true
167
  }
168
  ],
169
  "normalizer": null,
 
155
  "rstrip": false,
156
  "normalized": false,
157
  "special": true
 
 
 
 
 
 
 
 
 
158
  }
159
  ],
160
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -136,14 +136,6 @@
136
  "rstrip": false,
137
  "single_word": false,
138
  "special": true
139
- },
140
- "49152": {
141
- "content": "<pad>",
142
- "lstrip": false,
143
- "normalized": false,
144
- "rstrip": false,
145
- "single_word": false,
146
- "special": true
147
  }
148
  },
149
  "additional_special_tokens": [
@@ -155,7 +147,7 @@
155
  "clean_up_tokenization_spaces": false,
156
  "eos_token": "<|im_end|>",
157
  "model_max_length": 2048,
158
- "pad_token": "<pad>",
159
  "tokenizer_class": "GPT2Tokenizer",
160
  "unk_token": "<|endoftext|>",
161
  "vocab_size": 49152
 
136
  "rstrip": false,
137
  "single_word": false,
138
  "special": true
 
 
 
 
 
 
 
 
139
  }
140
  },
141
  "additional_special_tokens": [
 
147
  "clean_up_tokenization_spaces": false,
148
  "eos_token": "<|im_end|>",
149
  "model_max_length": 2048,
150
+ "pad_token": "<|im_end|>",
151
  "tokenizer_class": "GPT2Tokenizer",
152
  "unk_token": "<|endoftext|>",
153
  "vocab_size": 49152