HueyNemud
commited on
Commit
•
449ee92
1
Parent(s):
33467d0
Pretrained on pero-OCR, fine-tuned for NER on pero-OCR
Browse files- config.json +10 -10
- pytorch_model.bin +1 -1
- tokenizer.json +0 -0
- tokenizer_config.json +1 -1
- training_args.bin +2 -2
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"CamembertForTokenClassification"
|
5 |
],
|
@@ -17,22 +17,22 @@
|
|
17 |
"2": "I-PER",
|
18 |
"3": "I-MISC",
|
19 |
"4": "I-ORG",
|
20 |
-
"5": "I-
|
21 |
-
"6": "I-
|
22 |
-
"7": "I-
|
23 |
-
"8": "I-
|
24 |
},
|
25 |
"initializer_range": 0.02,
|
26 |
"intermediate_size": 3072,
|
27 |
"label2id": {
|
28 |
-
"I-ACT":
|
29 |
-
"I-CARDINAL":
|
30 |
-
"I-
|
31 |
"I-LOC": 1,
|
32 |
"I-MISC": 3,
|
33 |
"I-ORG": 4,
|
34 |
"I-PER": 2,
|
35 |
-
"I-TITRE":
|
36 |
"O": 0
|
37 |
},
|
38 |
"layer_norm_eps": 1e-05,
|
@@ -44,7 +44,7 @@
|
|
44 |
"pad_token_id": 1,
|
45 |
"position_embedding_type": "absolute",
|
46 |
"torch_dtype": "float32",
|
47 |
-
"transformers_version": "4.
|
48 |
"type_vocab_size": 1,
|
49 |
"use_cache": true,
|
50 |
"vocab_size": 32005
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "HueyNemud/das22-10-camembert_pretrained",
|
3 |
"architectures": [
|
4 |
"CamembertForTokenClassification"
|
5 |
],
|
|
|
17 |
"2": "I-PER",
|
18 |
"3": "I-MISC",
|
19 |
"4": "I-ORG",
|
20 |
+
"5": "I-CARDINAL",
|
21 |
+
"6": "I-ACT",
|
22 |
+
"7": "I-TITRE",
|
23 |
+
"8": "I-FT"
|
24 |
},
|
25 |
"initializer_range": 0.02,
|
26 |
"intermediate_size": 3072,
|
27 |
"label2id": {
|
28 |
+
"I-ACT": 6,
|
29 |
+
"I-CARDINAL": 5,
|
30 |
+
"I-FT": 8,
|
31 |
"I-LOC": 1,
|
32 |
"I-MISC": 3,
|
33 |
"I-ORG": 4,
|
34 |
"I-PER": 2,
|
35 |
+
"I-TITRE": 7,
|
36 |
"O": 0
|
37 |
},
|
38 |
"layer_norm_eps": 1e-05,
|
|
|
44 |
"pad_token_id": 1,
|
45 |
"position_embedding_type": "absolute",
|
46 |
"torch_dtype": "float32",
|
47 |
+
"transformers_version": "4.15.0",
|
48 |
"type_vocab_size": 1,
|
49 |
"use_cache": true,
|
50 |
"vocab_size": 32005
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 440237809
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10b4c449f459a1dd150bde11f8cf172d41be99a9f3b58800012ae433cc402cc2
|
3 |
size 440237809
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "additional_special_tokens": ["<s>NOTUSED", "</s>NOTUSED"], "model_max_length": 512, "name_or_path": "
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "additional_special_tokens": ["<s>NOTUSED", "</s>NOTUSED"], "model_max_length": 512, "name_or_path": "HueyNemud/das22-10-camembert_pretrained", "special_tokens_map_file": "/lrde/home2/jchazalo/.cache/huggingface/transformers/fe0e213c44079a9ee091098f81fff2941484006e9ba3001a9bf1ee9f87537599.cb3ec3a6c1200d181228d8825ae9767572abca54efa1bbb37fd83d721b2ef323", "sp_model_kwargs": {}, "tokenizer_class": "CamembertTokenizer"}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:31aeb7cf016b0e47597e331fa262b44873f81073521049d8a85dc46bd2004681
|
3 |
+
size 3055
|