guocheng98 commited on
Commit
17a96e1
1 Parent(s): 6bfd7c6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -7
README.md CHANGED
@@ -9,11 +9,9 @@ tags:
9
  license: apache-2.0
10
  ---
11
 
12
- # HelsinkiNLP-FineTuned-Legal-es-zh
13
-
14
  This model is a fine-tuned version of [Helsinki-NLP/opus-tatoeba-es-zh](https://huggingface.co/Helsinki-NLP/opus-tatoeba-es-zh) on a dataset of legal domain constructed by the author himself.
15
 
16
- ## Intended uses & limitations
17
 
18
  This model is the result of the master graduation thesis for the Tradumatics: Translation Technologies program at the Autonomous University of Barcelona.
19
 
@@ -21,13 +19,13 @@ Please refer to GitHub repo created for this thesis for full-text and relative o
21
 
22
  The thesis intends to explain various theories and certain algorithm details about neural machine translation, thus this fine-tuned model only serves as a hands-on practice example for that objective, without any intention of productive usage.
23
 
24
- ## Training and evaluation data
25
 
26
  The dataset is constructed from the Chinese translation of Spanish Civil Code, Spanish Constitution, and many other laws & regulations found in the database China Law Info (北大法宝 Beida Fabao), along with their source text found on Boletín Oficial del Estado and EUR-Lex.
27
 
28
  There are 9972 sentence pairs constructed. 1000 are used for evaluation and the rest for training.
29
 
30
- ## Training hyperparameters
31
 
32
  The following hyperparameters were used during training:
33
  - learning_rate: 2e-05
@@ -42,7 +40,7 @@ The following hyperparameters were used during training:
42
  - weight_decay: 0.01
43
  - early_stopping_patience: 8
44
 
45
- ## Training results
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
@@ -69,7 +67,7 @@ The following hyperparameters were used during training:
69
  | 1.1238 | 7.49 | 8400 | 2.1102 |
70
  | 1.1417 | 7.84 | 8800 | 2.1078 |
71
 
72
- ## Framework versions
73
 
74
  - Transformers 4.7.0
75
  - Pytorch 1.8.1+cu101
 
9
  license: apache-2.0
10
  ---
11
 
 
 
12
  This model is a fine-tuned version of [Helsinki-NLP/opus-tatoeba-es-zh](https://huggingface.co/Helsinki-NLP/opus-tatoeba-es-zh) on a dataset of legal domain constructed by the author himself.
13
 
14
+ # Intended uses & limitations
15
 
16
  This model is the result of the master graduation thesis for the Tradumatics: Translation Technologies program at the Autonomous University of Barcelona.
17
 
 
19
 
20
  The thesis intends to explain various theories and certain algorithm details about neural machine translation, thus this fine-tuned model only serves as a hands-on practice example for that objective, without any intention of productive usage.
21
 
22
+ # Training and evaluation data
23
 
24
  The dataset is constructed from the Chinese translation of Spanish Civil Code, Spanish Constitution, and many other laws & regulations found in the database China Law Info (北大法宝 Beida Fabao), along with their source text found on Boletín Oficial del Estado and EUR-Lex.
25
 
26
  There are 9972 sentence pairs constructed. 1000 are used for evaluation and the rest for training.
27
 
28
+ # Training hyperparameters
29
 
30
  The following hyperparameters were used during training:
31
  - learning_rate: 2e-05
 
40
  - weight_decay: 0.01
41
  - early_stopping_patience: 8
42
 
43
+ # Training results
44
 
45
  | Training Loss | Epoch | Step | Validation Loss |
46
  |:-------------:|:-----:|:----:|:---------------:|
 
67
  | 1.1238 | 7.49 | 8400 | 2.1102 |
68
  | 1.1417 | 7.84 | 8800 | 2.1078 |
69
 
70
+ # Framework versions
71
 
72
  - Transformers 4.7.0
73
  - Pytorch 1.8.1+cu101