murthyrudra commited on
Commit
f56dd59
1 Parent(s): ad2405a

End of training

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
GOAT ADDED
@@ -0,0 +1 @@
 
 
1
+ Subproject commit ad2405a8a4d57122bc11f88bcf129f7db4105d1e
config.json ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "../HiNER/cfilt_original/cfilt_original-muril-base-cased_ner_16_0.0001_10_1/",
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "embedding_size": 768,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "B-FESTIVAL",
14
+ "1": "B-GAME",
15
+ "2": "B-LANGUAGE",
16
+ "3": "B-LITERATURE",
17
+ "4": "B-LOCATION",
18
+ "5": "B-MISC",
19
+ "6": "B-NUMEX",
20
+ "7": "B-ORGANIZATION",
21
+ "8": "B-PERSON",
22
+ "9": "B-RELIGION",
23
+ "10": "B-TIMEX",
24
+ "11": "I-FESTIVAL",
25
+ "12": "I-GAME",
26
+ "13": "I-LANGUAGE",
27
+ "14": "I-LITERATURE",
28
+ "15": "I-LOCATION",
29
+ "16": "I-MISC",
30
+ "17": "I-NUMEX",
31
+ "18": "I-ORGANIZATION",
32
+ "19": "I-PERSON",
33
+ "20": "I-RELIGION",
34
+ "21": "I-TIMEX",
35
+ "22": "O"
36
+ },
37
+ "initializer_range": 0.02,
38
+ "intermediate_size": 3072,
39
+ "label2id": {
40
+ "B-FESTIVAL": 0,
41
+ "B-GAME": 1,
42
+ "B-LANGUAGE": 2,
43
+ "B-LITERATURE": 3,
44
+ "B-LOCATION": 4,
45
+ "B-MISC": 5,
46
+ "B-NUMEX": 6,
47
+ "B-ORGANIZATION": 7,
48
+ "B-PERSON": 8,
49
+ "B-RELIGION": 9,
50
+ "B-TIMEX": 10,
51
+ "I-FESTIVAL": 11,
52
+ "I-GAME": 12,
53
+ "I-LANGUAGE": 13,
54
+ "I-LITERATURE": 14,
55
+ "I-LOCATION": 15,
56
+ "I-MISC": 16,
57
+ "I-NUMEX": 17,
58
+ "I-ORGANIZATION": 18,
59
+ "I-PERSON": 19,
60
+ "I-RELIGION": 20,
61
+ "I-TIMEX": 21,
62
+ "O": 22
63
+ },
64
+ "layer_norm_eps": 1e-12,
65
+ "max_position_embeddings": 512,
66
+ "model_type": "bert",
67
+ "num_attention_heads": 12,
68
+ "num_hidden_layers": 12,
69
+ "pad_token_id": 0,
70
+ "position_embedding_type": "absolute",
71
+ "torch_dtype": "float32",
72
+ "transformers_version": "4.14.0",
73
+ "type_vocab_size": 2,
74
+ "use_cache": true,
75
+ "vocab_size": 197285
76
+ }
dev_predictions.txt ADDED
The diff for this file is too large to render. See raw diff
 
dev_results.txt ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ test_loss = 0.10530181974172592
2
+ test_precision = 0.8893083330733472
3
+ test_recall = 0.8804905170816087
4
+ test_f1 = 0.884877458208515
5
+ test_runtime = 126.3661
6
+ test_samples_per_second = 86.376
7
+ test_steps_per_second = 10.802
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:087be24cd1c2da2889b7d185b64351760ecb3a051ee8d2a6537d53548a0bad91
3
+ size 948016689
runs/May01_03-19-14_cccxc531/events.out.tfevents.1651390377.cccxc531.1595591.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:351d97c9096ccf3def98c584438ef1e77f3eb9d3c263808cc943b9d4f0dc7ae6
3
+ size 389
test_predictions.txt ADDED
The diff for this file is too large to render. See raw diff
 
test_results.txt ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ test_loss = 0.10449089109897614
2
+ test_precision = 0.8874067587220668
3
+ test_recall = 0.8801259383336435
4
+ test_f1 = 0.8837513529507954
5
+ test_runtime = 252.366
6
+ test_samples_per_second = 86.501
7
+ test_steps_per_second = 10.814
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a96f1e6331922276624faf193fe18c5f533f12ef36979181cb331bdcae589041
3
+ size 3055