anzorq commited on
Commit
56d96c8
1 Parent(s): 38984a6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -71
README.md CHANGED
@@ -1,72 +1,12 @@
1
- ---
2
  license: mit
3
- tags:
4
- - generated_from_trainer
5
- metrics:
6
- - bleu
7
- model-index:
8
- - name: m2m100_418M-finetuned-ko-to-en4
9
- results: []
10
- ---
11
-
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
-
15
- # m2m100_418M-finetuned-ko-to-en4
16
-
17
- This model is a fine-tuned version of [facebook/m2m100_418M](https://huggingface.co/facebook/m2m100_418M) on the None dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 0.4598
20
- - Bleu: 85.3745
21
- - Gen Len: 9.7522
22
-
23
- ## Model description
24
-
25
- More information needed
26
-
27
- ## Intended uses & limitations
28
-
29
- More information needed
30
-
31
- ## Training and evaluation data
32
-
33
- More information needed
34
-
35
- ## Training procedure
36
-
37
- ### Training hyperparameters
38
-
39
- The following hyperparameters were used during training:
40
- - learning_rate: 0.001
41
- - train_batch_size: 4
42
- - eval_batch_size: 4
43
- - seed: 42
44
- - gradient_accumulation_steps: 256
45
- - total_train_batch_size: 1024
46
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
- - lr_scheduler_type: linear
48
- - num_epochs: 10
49
- - mixed_precision_training: Native AMP
50
-
51
- ### Training results
52
-
53
- | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
54
- |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
55
- | No log | 1.0 | 105 | 1.8667 | 24.5072 | 9.523 |
56
- | No log | 2.0 | 210 | 0.8581 | 57.9973 | 9.2779 |
57
- | No log | 3.0 | 315 | 0.6587 | 69.4588 | 9.7399 |
58
- | No log | 4.0 | 420 | 0.5762 | 74.5636 | 9.6775 |
59
- | 1.4539 | 5.0 | 525 | 0.5254 | 78.8897 | 9.6946 |
60
- | 1.4539 | 6.0 | 630 | 0.4952 | 81.0054 | 9.7073 |
61
- | 1.4539 | 7.0 | 735 | 0.4773 | 83.0792 | 9.7233 |
62
- | 1.4539 | 8.0 | 840 | 0.4669 | 84.4309 | 9.7429 |
63
- | 1.4539 | 9.0 | 945 | 0.4616 | 85.0965 | 9.749 |
64
- | 0.144 | 10.0 | 1050 | 0.4598 | 85.3745 | 9.7522 |
65
-
66
-
67
- ### Framework versions
68
-
69
- - Transformers 4.18.0
70
- - Pytorch 1.11.0+cu113
71
- - Datasets 2.1.0
72
- - Tokenizers 0.12.1
 
 
1
  license: mit
2
+ language:
3
+ - ru
4
+ - kbd
5
+ widget:
6
+ - text: "Я иду домой."
7
+ example_title: "Я иду домой."
8
+ - text: "Дети играют во дворе."
9
+ example_title: "Дети играют во дворе."
10
+ - text: "Сколько тебе лет?"
11
+ example_title: "Сколько тебе лет?"
12
+ ---