danielsteinigen commited on
Commit
ac775f9
1 Parent(s): 5c7b467

Upload tokenizer

Browse files
.gitattributes CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "[/ENT-01]": 250002,
3
+ "[/ENT-02]": 250003,
4
+ "[/ENT-03]": 250004,
5
+ "[/ENT-04]": 250005,
6
+ "[/ENT-05]": 250006,
7
+ "[/ENT-06]": 250007,
8
+ "[/ENT-07]": 250008,
9
+ "[/ENT-08]": 250009,
10
+ "[/OBJ]": 250010,
11
+ "[/SUB]": 250011,
12
+ "[ENT-01]": 250012,
13
+ "[ENT-02]": 250013,
14
+ "[ENT-03]": 250014,
15
+ "[ENT-04]": 250015,
16
+ "[ENT-05]": 250016,
17
+ "[ENT-06]": 250017,
18
+ "[ENT-07]": 250018,
19
+ "[ENT-08]": 250019,
20
+ "[GRP-00]": 250020,
21
+ "[GRP-01]": 250021,
22
+ "[OBJ]": 250022,
23
+ "[REL]": 250023,
24
+ "[SUB]": 250024
25
+ }
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
special_tokens_map.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "[/ENT-01]",
4
+ "[/ENT-02]",
5
+ "[/ENT-03]",
6
+ "[/ENT-04]",
7
+ "[/ENT-05]",
8
+ "[/ENT-06]",
9
+ "[/ENT-07]",
10
+ "[/ENT-08]",
11
+ "[/OBJ]",
12
+ "[/SUB]",
13
+ "[ENT-01]",
14
+ "[ENT-02]",
15
+ "[ENT-03]",
16
+ "[ENT-04]",
17
+ "[ENT-05]",
18
+ "[ENT-06]",
19
+ "[ENT-07]",
20
+ "[ENT-08]",
21
+ "[GRP-00]",
22
+ "[GRP-01]",
23
+ "[OBJ]",
24
+ "[REL]",
25
+ "[SUB]"
26
+ ],
27
+ "bos_token": "<s>",
28
+ "cls_token": "<s>",
29
+ "eos_token": "</s>",
30
+ "mask_token": {
31
+ "content": "<mask>",
32
+ "lstrip": true,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "pad_token": "<pad>",
38
+ "sep_token": "</s>",
39
+ "unk_token": "<unk>"
40
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f6c6cd2187fd44407ae578446f51e29aef61147368380b76f051a6138cb41d5
3
+ size 17086910
tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "__type": "AddedToken",
7
+ "content": "<mask>",
8
+ "lstrip": true,
9
+ "normalized": true,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "max_length": 512,
14
+ "model_max_length": 512,
15
+ "name_or_path": "xlm-roberta-large",
16
+ "pad_token": "<pad>",
17
+ "padding": "max_length",
18
+ "sep_token": "</s>",
19
+ "special_tokens_map_file": null,
20
+ "tokenizer_class": "XLMRobertaTokenizer",
21
+ "truncation": true,
22
+ "unk_token": "<unk>"
23
+ }