Upload tokenizer
Browse files- tokenizer.json +2 -2
- tokenizer_config.json +0 -4
tokenizer.json
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d842e6af904403ce6bf8ee58faffd9abad1682513c28c27454d81dc67eaf296c
|
3 |
+
size 16315149
|
tokenizer_config.json
CHANGED
@@ -30,13 +30,9 @@
|
|
30 |
"eos_token": "</s>",
|
31 |
"extra_ids": 0,
|
32 |
"legacy": true,
|
33 |
-
"max_length": 512,
|
34 |
"model_max_length": 1000000000000000019884624838656,
|
35 |
"pad_token": "<pad>",
|
36 |
"sp_model_kwargs": {},
|
37 |
-
"stride": 0,
|
38 |
"tokenizer_class": "T5Tokenizer",
|
39 |
-
"truncation_side": "right",
|
40 |
-
"truncation_strategy": "longest_first",
|
41 |
"unk_token": "<unk>"
|
42 |
}
|
|
|
30 |
"eos_token": "</s>",
|
31 |
"extra_ids": 0,
|
32 |
"legacy": true,
|
|
|
33 |
"model_max_length": 1000000000000000019884624838656,
|
34 |
"pad_token": "<pad>",
|
35 |
"sp_model_kwargs": {},
|
|
|
36 |
"tokenizer_class": "T5Tokenizer",
|
|
|
|
|
37 |
"unk_token": "<unk>"
|
38 |
}
|