BioBERT V1.1 model update
Browse files- README.md +136 -0
- config.json +62 -62
- optimizer.pt +1 -1
- pytorch_model.bin +1 -1
- rng_state.pth +2 -2
- scheduler.pt +1 -1
- trainer_state.json +48 -84
- training_args.bin +1 -1
README.md
CHANGED
@@ -1,3 +1,139 @@
|
|
1 |
---
|
|
|
|
|
|
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
language: fr
|
3 |
+
datasets:
|
4 |
+
- FrenchMedMCQA
|
5 |
license: apache-2.0
|
6 |
+
model-index:
|
7 |
+
- name: qanastek/FrenchMedMCQA-BioBERT-V1.1-Wikipedia-BM25
|
8 |
+
results:
|
9 |
+
- task:
|
10 |
+
type: question-answering
|
11 |
+
name: Question Answering
|
12 |
+
dataset:
|
13 |
+
name: FrenchMedMCQA
|
14 |
+
type: FrenchMedMCQA
|
15 |
+
config: FrenchMedMCQA
|
16 |
+
split: validation
|
17 |
+
metrics:
|
18 |
+
- name: Exact Match
|
19 |
+
type: exact_match
|
20 |
+
value: 16.72
|
21 |
+
verified: true
|
22 |
+
- name: Hamming Score
|
23 |
+
type: hamming score
|
24 |
+
value: 38.72
|
25 |
+
verified: true
|
26 |
+
widget:
|
27 |
+
- text: "Parmi les affirmations suivantes, une seule est fausse, indiquer laquelle: les particules alpha \n (A) Sont formées de noyaux d'hélium (B) Sont peu pénétrantes (C) Toute l'énergie qu'elles transportent est cédée au long d'un parcours de quelques centimètres dans l'air (D) Sont arrêtées par une feuille de papier (E) Sont peu ionisantes\n Acception particulière"
|
28 |
---
|
29 |
+
|
30 |
+
# FrenchMedMCQA : Multiple-choice question answering on pharmacology exams using BioBERT V1.1, Wikipedia external knowledge and BM25 retriever
|
31 |
+
|
32 |
+
- Corpora: [FrenchMedMCQA](https://github.com/qanastek/FrenchMedMCQA)
|
33 |
+
- Model: [BioBERT V1.1](https://huggingface.co/dmis-lab/biobert-v1.1)
|
34 |
+
- Number of Epochs: 10
|
35 |
+
|
36 |
+
**People Involved**
|
37 |
+
|
38 |
+
* [Yanis LABRAK](https://www.linkedin.com/in/yanis-labrak-8a7412145/) (1)
|
39 |
+
* [Adrien BAZOGE](https://fr.linkedin.com/in/adrien-bazoge-6b511b145) (2)
|
40 |
+
* [Richard DUFOUR](https://cv.archives-ouvertes.fr/richard-dufour) (2)
|
41 |
+
* [Béatrice DAILLE](https://scholar.google.com/citations?user=-damXYEAAAAJ&hl=fr) (2)
|
42 |
+
* [Pierre-Antoine GOURRAUD](https://fr.linkedin.com/in/pierre-antoine-gourraud-35779b6) (3)
|
43 |
+
* [Emmanuel MORIN](https://scholar.google.fr/citations?user=tvTEtM0AAAAJ&hl=fr) (2)
|
44 |
+
* [Mickael ROUVIER](https://scholar.google.fr/citations?user=0fmu-VsAAAAJ&hl=fr) (1)
|
45 |
+
|
46 |
+
**Affiliations**
|
47 |
+
|
48 |
+
1. [LIA, NLP team](https://lia.univ-avignon.fr/), Avignon University, Avignon, France.
|
49 |
+
2. [LS2N, TALN team](https://www.ls2n.fr/equipe/taln/), Nantes University, Nantes, France.
|
50 |
+
3. [CHU Nantes](https://www.chu-nantes.fr/), Nantes University, Nantes, France.
|
51 |
+
|
52 |
+
## Demo: How to use in HuggingFace Transformers
|
53 |
+
|
54 |
+
Requires [Transformers](https://pypi.org/project/transformers/): ```pip install transformers```
|
55 |
+
|
56 |
+
```python
|
57 |
+
from datasets import load_dataset
|
58 |
+
from transformers import AutoTokenizer, AutoModelForSequenceClassification, TextClassificationPipeline
|
59 |
+
|
60 |
+
path_model = "qanastek/FrenchMedMCQA-BioBERT-V1.1-Wikipedia-BM25"
|
61 |
+
|
62 |
+
tokenizer = AutoTokenizer.from_pretrained(path_model)
|
63 |
+
model = AutoModelForSequenceClassification.from_pretrained(path_model)
|
64 |
+
|
65 |
+
pipeline = TextClassificationPipeline(model=model, tokenizer=tokenizer, return_all_scores=False, device=0) # GPU
|
66 |
+
|
67 |
+
dataset = load_dataset("qanastek/FrenchMedMCQA")["test"]
|
68 |
+
|
69 |
+
for e in dataset:
|
70 |
+
prediction = pipeline(e["bert_text"], truncation=True, max_length=model.config.max_position_embeddings)
|
71 |
+
```
|
72 |
+
|
73 |
+
Output:
|
74 |
+
|
75 |
+
![Preview Output](preview.PNG)
|
76 |
+
|
77 |
+
## Training data
|
78 |
+
|
79 |
+
The questions and their associated candidate answer(s) were collected from real French pharmacy exams on the remede website. Questions and answers were manually created by medical experts and used during examinations. The dataset is composed of 2,025 questions with multiple answers and 1,080 with a single one, for a total of 3,105 questions. Each instance of the dataset contains an identifier, a question, five options (labeled from A to E) and correct answer(s). The average question length is 14.17 tokens and the average answer length is 6.44 tokens. The vocabulary size is of 13k words, of which 3.8k are estimated medical domain-specific words (i.e. a word related to the medical field). We find an average of 2.49 medical domain-specific words in each question (17 % of the words) and 2 in each answer (36 % of the words). On average, a medical domain-specific word is present in 2 questions and in 8 answers.
|
80 |
+
|
81 |
+
| # Answers | Training | Validation | Test | Total |
|
82 |
+
|:---------:|:--------:|:----------:|:----:|:-----:|
|
83 |
+
| 1 | 595 | 164 | 321 | 1,080 |
|
84 |
+
| 2 | 528 | 45 | 97 | 670 |
|
85 |
+
| 3 | 718 | 71 | 141 | 930 |
|
86 |
+
| 4 | 296 | 30 | 56 | 382 |
|
87 |
+
| 5 | 34 | 2 | 7 | 43 |
|
88 |
+
| Total | 2171 | 312 | 622 | 3,105 |
|
89 |
+
|
90 |
+
## Evaluation results
|
91 |
+
|
92 |
+
The test corpora used for this evaluation is available on [Github](https://github.com/qanastek/FrenchMedMCQA).
|
93 |
+
|
94 |
+
| Architecture | Hamming | EMR | Hamming | EMR | Hamming | EMR | Hamming | EMR | Hamming | EMR |
|
95 |
+
|:----------------:|:-------:|:-----:|:-------:|:-----:|:-------:|:-----:|:-------:|:-----:|:-------:|:-----:|
|
96 |
+
| BioBERT V1.1 | 36.19 | 15.43 | **38.72** | 16.72 | 33.33 | 14.14 | 35.13 | 16.23 | 34.27 | 13.98 |
|
97 |
+
| PubMedBERT | 33.98 | 14.14 | 34.00 | 13.98 | 35.66 | 15.59 | 33.87 | 14.79 | 35.44 | 14.79 |
|
98 |
+
| CamemBERT-base | 36.24 | 16.55 | 34.19 | 14.46 | 34.78 | 15.43 | 34.66 | 14.79 | 34.61 | 14.95 |
|
99 |
+
| XLM-RoBERTa-base | 37.92 | 17.20 | 31.26 | 11.89 | 35.84 | 16.07 | 32.47 | 14.63 | 33.00 | 14.95 |
|
100 |
+
| BART-base | 31.93 | 15.91 | 34.98 | **18.64** | 33.80 | 17.68 | 29.65 | 12.86 | 34.65 | 18.32 |
|
101 |
+
|
102 |
+
## BibTeX Citations
|
103 |
+
|
104 |
+
Please cite the following paper when using this model.
|
105 |
+
|
106 |
+
FrenchMedMCQA corpus and linked tools:
|
107 |
+
|
108 |
+
```latex
|
109 |
+
@unpublished{labrak:hal-03824241,
|
110 |
+
TITLE = {{FrenchMedMCQA: A French Multiple-Choice Question Answering Dataset for Medical domain}},
|
111 |
+
AUTHOR = {Labrak, Yanis and Bazoge, Adrien and Dufour, Richard and Daille, B{\'e}atrice and Gourraud, Pierre-Antoine and Morin, Emmanuel and Rouvier, Mickael},
|
112 |
+
URL = {https://hal.archives-ouvertes.fr/hal-03824241},
|
113 |
+
NOTE = {working paper or preprint},
|
114 |
+
YEAR = {2022},
|
115 |
+
MONTH = Oct,
|
116 |
+
PDF = {https://hal.archives-ouvertes.fr/hal-03824241/file/LOUHI_2022___QA-3.pdf},
|
117 |
+
HAL_ID = {hal-03824241},
|
118 |
+
HAL_VERSION = {v1},
|
119 |
+
}
|
120 |
+
```
|
121 |
+
|
122 |
+
HuggingFace's Transformers :
|
123 |
+
|
124 |
+
```latex
|
125 |
+
@misc{https://doi.org/10.48550/arxiv.1910.03771,
|
126 |
+
doi = {10.48550/ARXIV.1910.03771},
|
127 |
+
url = {https://arxiv.org/abs/1910.03771},
|
128 |
+
author = {Wolf, Thomas and Debut, Lysandre and Sanh, Victor and Chaumond, Julien and Delangue, Clement and Moi, Anthony and Cistac, Pierric and Rault, Tim and Louf, Rémi and Funtowicz, Morgan and Davison, Joe and Shleifer, Sam and von Platen, Patrick and Ma, Clara and Jernite, Yacine and Plu, Julien and Xu, Canwen and Scao, Teven Le and Gugger, Sylvain and Drame, Mariama and Lhoest, Quentin and Rush, Alexander M.},
|
129 |
+
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
|
130 |
+
title = {HuggingFace's Transformers: State-of-the-art Natural Language Processing},
|
131 |
+
publisher = {arXiv},
|
132 |
+
year = {2019},
|
133 |
+
copyright = {arXiv.org perpetual, non-exclusive license}
|
134 |
+
}
|
135 |
+
```
|
136 |
+
|
137 |
+
## Acknowledgment
|
138 |
+
|
139 |
+
This work was financially supported by [Zenidoc](https://zenidoc.fr/), the [DIETS](https://anr-diets.univ-avignon.fr/) project financed by the Agence Nationale de la Recherche (ANR) under contract ANR-20-CE23-0005 and the ANR [AIBy4](https://aiby4.ls2n.fr/) (ANR-20-THIA-0011).
|
config.json
CHANGED
@@ -10,72 +10,72 @@
|
|
10 |
"hidden_dropout_prob": 0.1,
|
11 |
"hidden_size": 768,
|
12 |
"id2label": {
|
13 |
-
"0": "
|
14 |
-
"1": "
|
15 |
-
"2": "
|
16 |
-
"3": "
|
17 |
-
"4": "
|
18 |
-
"5": "
|
19 |
-
"6": "
|
20 |
-
"7": "
|
21 |
-
"8": "
|
22 |
-
"9": "
|
23 |
-
"10": "
|
24 |
-
"11": "
|
25 |
-
"12": "
|
26 |
-
"13": "
|
27 |
-
"14": "
|
28 |
-
"15": "
|
29 |
-
"16": "
|
30 |
-
"17": "
|
31 |
-
"18": "
|
32 |
-
"19": "
|
33 |
-
"20": "
|
34 |
-
"21": "
|
35 |
-
"22": "
|
36 |
-
"23": "
|
37 |
-
"24": "
|
38 |
-
"25": "
|
39 |
-
"26": "
|
40 |
-
"27": "
|
41 |
-
"28": "
|
42 |
-
"29": "
|
43 |
-
"30": "
|
44 |
},
|
45 |
"initializer_range": 0.02,
|
46 |
"intermediate_size": 3072,
|
47 |
"label2id": {
|
48 |
-
"
|
49 |
-
"
|
50 |
-
"
|
51 |
-
"
|
52 |
-
"
|
53 |
-
"
|
54 |
-
"
|
55 |
-
"
|
56 |
-
"
|
57 |
-
"
|
58 |
-
"
|
59 |
-
"
|
60 |
-
"
|
61 |
-
"
|
62 |
-
"
|
63 |
-
"
|
64 |
-
"
|
65 |
-
"
|
66 |
-
"
|
67 |
-
"
|
68 |
-
"
|
69 |
-
"
|
70 |
-
"
|
71 |
-
"
|
72 |
-
"
|
73 |
-
"
|
74 |
-
"
|
75 |
-
"
|
76 |
-
"
|
77 |
-
"
|
78 |
-
"
|
79 |
},
|
80 |
"layer_norm_eps": 1e-12,
|
81 |
"max_position_embeddings": 512,
|
|
|
10 |
"hidden_dropout_prob": 0.1,
|
11 |
"hidden_size": 768,
|
12 |
"id2label": {
|
13 |
+
"0": "c",
|
14 |
+
"1": "a",
|
15 |
+
"2": "e",
|
16 |
+
"3": "d",
|
17 |
+
"4": "b",
|
18 |
+
"5": "be",
|
19 |
+
"6": "ae",
|
20 |
+
"7": "bc",
|
21 |
+
"8": "bd",
|
22 |
+
"9": "ab",
|
23 |
+
"10": "de",
|
24 |
+
"11": "cd",
|
25 |
+
"12": "ac",
|
26 |
+
"13": "ad",
|
27 |
+
"14": "ce",
|
28 |
+
"15": "bce",
|
29 |
+
"16": "abc",
|
30 |
+
"17": "cde",
|
31 |
+
"18": "bcd",
|
32 |
+
"19": "ace",
|
33 |
+
"20": "ade",
|
34 |
+
"21": "abe",
|
35 |
+
"22": "acd",
|
36 |
+
"23": "bde",
|
37 |
+
"24": "abd",
|
38 |
+
"25": "abde",
|
39 |
+
"26": "abcd",
|
40 |
+
"27": "bcde",
|
41 |
+
"28": "abce",
|
42 |
+
"29": "acde",
|
43 |
+
"30": "abcde"
|
44 |
},
|
45 |
"initializer_range": 0.02,
|
46 |
"intermediate_size": 3072,
|
47 |
"label2id": {
|
48 |
+
"c": 0,
|
49 |
+
"a": 1,
|
50 |
+
"e": 10,
|
51 |
+
"d": 11,
|
52 |
+
"b": 12,
|
53 |
+
"be": 13,
|
54 |
+
"ae": 14,
|
55 |
+
"bc": 15,
|
56 |
+
"bd": 16,
|
57 |
+
"ab": 17,
|
58 |
+
"de": 18,
|
59 |
+
"cd": 19,
|
60 |
+
"ac": 2,
|
61 |
+
"ad": 20,
|
62 |
+
"ce": 21,
|
63 |
+
"bce": 22,
|
64 |
+
"abc": 23,
|
65 |
+
"cde": 24,
|
66 |
+
"bcd": 25,
|
67 |
+
"ace": 26,
|
68 |
+
"ade": 27,
|
69 |
+
"abe": 28,
|
70 |
+
"acd": 29,
|
71 |
+
"bde": 3,
|
72 |
+
"abd": 30,
|
73 |
+
"abde": 4,
|
74 |
+
"abcd": 5,
|
75 |
+
"bcde": 6,
|
76 |
+
"abce": 7,
|
77 |
+
"acde": 8,
|
78 |
+
"abcde": 9
|
79 |
},
|
80 |
"layer_norm_eps": 1e-12,
|
81 |
"max_position_embeddings": 512,
|
optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 866789277
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:233cf2771eb01d2e8c73fcc050e67de43b98b97246fd7b17a452cd8b633619aa
|
3 |
size 866789277
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 433407405
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba1c8d0fe5c7d27a9fc6cf9470f544dd35775450cd60e42b157b06e200e97e8b
|
3 |
size 433407405
|
rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee20a722b64f81302a306cfea71de1f48196838ef4a0215073c03138fafe0f41
|
3 |
+
size 14503
|
scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:229af39592678ac24e5cdc1157fbb6b2c3e0d09815db0c4f615dd862058567fd
|
3 |
size 623
|
trainer_state.json
CHANGED
@@ -1,106 +1,70 @@
|
|
1 |
{
|
2 |
-
"best_metric": 0.
|
3 |
-
"best_model_checkpoint": "biobert-v1.1-finetuned-frenchmedmcqa/checkpoint-
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
9 |
"log_history": [
|
10 |
{
|
11 |
-
"epoch":
|
12 |
-
"
|
13 |
-
"
|
14 |
-
"
|
15 |
-
"eval_precision": 0.01899449375410914,
|
16 |
-
"eval_recall": 0.13782051282051283,
|
17 |
-
"eval_runtime": 5.2204,
|
18 |
-
"eval_samples_per_second": 59.765,
|
19 |
-
"eval_steps_per_second": 3.831,
|
20 |
-
"step": 136
|
21 |
-
},
|
22 |
-
{
|
23 |
-
"epoch": 2.0,
|
24 |
-
"eval_accuracy": 0.16025641025641027,
|
25 |
-
"eval_f1": 0.06942300892072582,
|
26 |
-
"eval_loss": 2.8957390785217285,
|
27 |
-
"eval_precision": 0.04520268867434859,
|
28 |
-
"eval_recall": 0.16025641025641027,
|
29 |
-
"eval_runtime": 5.1606,
|
30 |
-
"eval_samples_per_second": 60.458,
|
31 |
-
"eval_steps_per_second": 3.875,
|
32 |
-
"step": 272
|
33 |
-
},
|
34 |
-
{
|
35 |
-
"epoch": 3.0,
|
36 |
-
"eval_accuracy": 0.15384615384615385,
|
37 |
-
"eval_f1": 0.0817681987234755,
|
38 |
-
"eval_loss": 2.722668170928955,
|
39 |
-
"eval_precision": 0.10610228539576365,
|
40 |
-
"eval_recall": 0.15384615384615385,
|
41 |
-
"eval_runtime": 5.1585,
|
42 |
-
"eval_samples_per_second": 60.483,
|
43 |
-
"eval_steps_per_second": 3.877,
|
44 |
-
"step": 408
|
45 |
},
|
46 |
{
|
47 |
-
"epoch":
|
48 |
-
"
|
49 |
-
"
|
50 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
51 |
},
|
52 |
{
|
53 |
-
"epoch":
|
54 |
-
"
|
55 |
-
"
|
56 |
-
"
|
57 |
-
"eval_precision": 0.061101876500489406,
|
58 |
-
"eval_recall": 0.15705128205128205,
|
59 |
-
"eval_runtime": 5.1556,
|
60 |
-
"eval_samples_per_second": 60.516,
|
61 |
-
"eval_steps_per_second": 3.879,
|
62 |
-
"step": 544
|
63 |
},
|
64 |
{
|
65 |
-
"epoch":
|
66 |
-
"eval_accuracy": 0.
|
67 |
-
"eval_f1": 0.
|
68 |
-
"eval_loss": 2.
|
69 |
-
"eval_precision": 0.
|
70 |
-
"eval_recall": 0.
|
71 |
-
"eval_runtime":
|
72 |
-
"eval_samples_per_second":
|
73 |
-
"eval_steps_per_second":
|
74 |
-
"step":
|
75 |
},
|
76 |
{
|
77 |
-
"epoch":
|
78 |
-
"
|
79 |
-
"
|
80 |
-
"
|
81 |
-
"eval_precision": 0.08323137866407097,
|
82 |
-
"eval_recall": 0.13782051282051283,
|
83 |
-
"eval_runtime": 5.1788,
|
84 |
-
"eval_samples_per_second": 60.246,
|
85 |
-
"eval_steps_per_second": 3.862,
|
86 |
-
"step": 816
|
87 |
},
|
88 |
{
|
89 |
-
"epoch":
|
90 |
-
"eval_accuracy": 0.
|
91 |
-
"eval_f1": 0.
|
92 |
-
"eval_loss": 2.
|
93 |
-
"eval_precision": 0.
|
94 |
-
"eval_recall": 0.
|
95 |
-
"eval_runtime":
|
96 |
-
"eval_samples_per_second":
|
97 |
-
"eval_steps_per_second":
|
98 |
-
"step":
|
99 |
}
|
100 |
],
|
101 |
-
"max_steps":
|
102 |
"num_train_epochs": 10,
|
103 |
-
"total_flos":
|
104 |
"trial_name": null,
|
105 |
"trial_params": null
|
106 |
}
|
|
|
1 |
{
|
2 |
+
"best_metric": 0.18269230769230768,
|
3 |
+
"best_model_checkpoint": "biobert-v1.1-finetuned-frenchmedmcqa/checkpoint-1629",
|
4 |
+
"epoch": 3.0,
|
5 |
+
"global_step": 1629,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
9 |
"log_history": [
|
10 |
{
|
11 |
+
"epoch": 0.92,
|
12 |
+
"learning_rate": 1.815837937384899e-05,
|
13 |
+
"loss": 3.4112,
|
14 |
+
"step": 500
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
},
|
16 |
{
|
17 |
+
"epoch": 1.0,
|
18 |
+
"eval_accuracy": 0.14423076923076922,
|
19 |
+
"eval_f1": 0.04993107948113005,
|
20 |
+
"eval_loss": 2.976611614227295,
|
21 |
+
"eval_precision": 0.030627588098176332,
|
22 |
+
"eval_recall": 0.14423076923076922,
|
23 |
+
"eval_runtime": 1.4297,
|
24 |
+
"eval_samples_per_second": 218.234,
|
25 |
+
"eval_steps_per_second": 54.559,
|
26 |
+
"step": 543
|
27 |
},
|
28 |
{
|
29 |
+
"epoch": 1.84,
|
30 |
+
"learning_rate": 1.6316758747697976e-05,
|
31 |
+
"loss": 3.1756,
|
32 |
+
"step": 1000
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
},
|
34 |
{
|
35 |
+
"epoch": 2.0,
|
36 |
+
"eval_accuracy": 0.1794871794871795,
|
37 |
+
"eval_f1": 0.0727026201206936,
|
38 |
+
"eval_loss": 2.687720775604248,
|
39 |
+
"eval_precision": 0.04664191343365827,
|
40 |
+
"eval_recall": 0.1794871794871795,
|
41 |
+
"eval_runtime": 1.3011,
|
42 |
+
"eval_samples_per_second": 239.802,
|
43 |
+
"eval_steps_per_second": 59.95,
|
44 |
+
"step": 1086
|
45 |
},
|
46 |
{
|
47 |
+
"epoch": 2.76,
|
48 |
+
"learning_rate": 1.4475138121546963e-05,
|
49 |
+
"loss": 3.0228,
|
50 |
+
"step": 1500
|
|
|
|
|
|
|
|
|
|
|
|
|
51 |
},
|
52 |
{
|
53 |
+
"epoch": 3.0,
|
54 |
+
"eval_accuracy": 0.18269230769230768,
|
55 |
+
"eval_f1": 0.11557929420998879,
|
56 |
+
"eval_loss": 2.6114799976348877,
|
57 |
+
"eval_precision": 0.10303806777379766,
|
58 |
+
"eval_recall": 0.18269230769230768,
|
59 |
+
"eval_runtime": 1.0933,
|
60 |
+
"eval_samples_per_second": 285.38,
|
61 |
+
"eval_steps_per_second": 71.345,
|
62 |
+
"step": 1629
|
63 |
}
|
64 |
],
|
65 |
+
"max_steps": 5430,
|
66 |
"num_train_epochs": 10,
|
67 |
+
"total_flos": 664497557825640.0,
|
68 |
"trial_name": null,
|
69 |
"trial_params": null
|
70 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3119
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5839cab219119544ad1ab8c8dec230e90c1e0fcc001fc00325b44b92acc9ba71
|
3 |
size 3119
|