edchengg
commited on
Commit
•
2d9c74e
1
Parent(s):
540b2bd
Update readme
Browse files
README.md
CHANGED
@@ -1,55 +1,271 @@
|
|
1 |
---
|
2 |
language:
|
3 |
-
-
|
4 |
-
-
|
5 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
tags:
|
7 |
-
-
|
8 |
-
|
9 |
-
|
10 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
14 |
should probably proofread and complete it, then remove this comment. -->
|
15 |
|
16 |
-
|
17 |
-
|
18 |
-
This model is a fine-tuned version of [facebook/nllb-200-3.3B](https://huggingface.co/facebook/nllb-200-3.3B) on the None dataset.
|
19 |
-
|
20 |
-
## Model description
|
21 |
|
22 |
-
|
23 |
-
|
24 |
-
## Intended uses & limitations
|
25 |
-
|
26 |
-
More information needed
|
27 |
|
28 |
-
|
|
|
|
|
|
|
29 |
|
30 |
-
|
|
|
|
|
31 |
|
32 |
-
|
|
|
|
|
|
|
33 |
|
34 |
-
|
|
|
|
|
|
|
|
|
|
|
35 |
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
- eval_batch_size: 1
|
40 |
-
- seed: 42
|
41 |
-
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
42 |
-
- lr_scheduler_type: linear
|
43 |
-
- lr_scheduler_warmup_steps: 100
|
44 |
-
- training_steps: 1000
|
45 |
|
46 |
-
|
|
|
|
|
|
|
|
|
47 |
|
|
|
|
|
|
|
48 |
|
|
|
|
|
|
|
|
|
49 |
|
50 |
-
|
51 |
|
52 |
-
|
53 |
-
|
54 |
-
-
|
55 |
-
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
language:
|
3 |
+
- ace
|
4 |
+
- acm
|
5 |
+
- acq
|
6 |
+
- aeb
|
7 |
+
- af
|
8 |
+
- ajp
|
9 |
+
- ak
|
10 |
+
- als
|
11 |
+
- am
|
12 |
+
- apc
|
13 |
+
- ar
|
14 |
+
- ars
|
15 |
+
- ary
|
16 |
+
- arz
|
17 |
+
- as
|
18 |
+
- ast
|
19 |
+
- awa
|
20 |
+
- ayr
|
21 |
+
- azb
|
22 |
+
- azj
|
23 |
+
- ba
|
24 |
+
- bm
|
25 |
+
- ban
|
26 |
+
- be
|
27 |
+
- bem
|
28 |
+
- bn
|
29 |
+
- bho
|
30 |
+
- bjn
|
31 |
+
- bo
|
32 |
+
- bs
|
33 |
+
- bug
|
34 |
+
- bg
|
35 |
+
- ca
|
36 |
+
- ceb
|
37 |
+
- cs
|
38 |
+
- cjk
|
39 |
+
- ckb
|
40 |
+
- crh
|
41 |
+
- cy
|
42 |
+
- da
|
43 |
+
- de
|
44 |
+
- dik
|
45 |
+
- dyu
|
46 |
+
- dz
|
47 |
+
- el
|
48 |
+
- en
|
49 |
+
- eo
|
50 |
+
- et
|
51 |
+
- eu
|
52 |
+
- ee
|
53 |
+
- fo
|
54 |
+
- fj
|
55 |
+
- fi
|
56 |
+
- fon
|
57 |
+
- fr
|
58 |
+
- fur
|
59 |
+
- fuv
|
60 |
+
- gaz
|
61 |
+
- gd
|
62 |
+
- ga
|
63 |
+
- gl
|
64 |
+
- gn
|
65 |
+
- gu
|
66 |
+
- ht
|
67 |
+
- ha
|
68 |
+
- he
|
69 |
+
- hi
|
70 |
+
- hne
|
71 |
+
- hr
|
72 |
+
- hu
|
73 |
+
- hy
|
74 |
+
- ig
|
75 |
+
- ilo
|
76 |
+
- id
|
77 |
+
- is
|
78 |
+
- it
|
79 |
+
- jv
|
80 |
+
- ja
|
81 |
+
- kab
|
82 |
+
- kac
|
83 |
+
- kam
|
84 |
+
- kn
|
85 |
+
- ks
|
86 |
+
- ka
|
87 |
+
- kk
|
88 |
+
- kbp
|
89 |
+
- kea
|
90 |
+
- khk
|
91 |
+
- km
|
92 |
+
- ki
|
93 |
+
- rw
|
94 |
+
- ky
|
95 |
+
- kmb
|
96 |
+
- kmr
|
97 |
+
- knc
|
98 |
+
- kg
|
99 |
+
- ko
|
100 |
+
- lo
|
101 |
+
- lij
|
102 |
+
- li
|
103 |
+
- ln
|
104 |
+
- lt
|
105 |
+
- lmo
|
106 |
+
- ltg
|
107 |
+
- lb
|
108 |
+
- lua
|
109 |
+
- lg
|
110 |
+
- luo
|
111 |
+
- lus
|
112 |
+
- lvs
|
113 |
+
- mag
|
114 |
+
- mai
|
115 |
+
- ml
|
116 |
+
- mar
|
117 |
+
- min
|
118 |
+
- mk
|
119 |
+
- mt
|
120 |
+
- mni
|
121 |
+
- mos
|
122 |
+
- mi
|
123 |
+
- my
|
124 |
+
- nl
|
125 |
+
- nn
|
126 |
+
- nb
|
127 |
+
- npi
|
128 |
+
- nso
|
129 |
+
- nus
|
130 |
+
- ny
|
131 |
+
- oc
|
132 |
+
- ory
|
133 |
+
- pag
|
134 |
+
- pa
|
135 |
+
- pap
|
136 |
+
- pbt
|
137 |
+
- pes
|
138 |
+
- plt
|
139 |
+
- pl
|
140 |
+
- pt
|
141 |
+
- prs
|
142 |
+
- quy
|
143 |
+
- ro
|
144 |
+
- rn
|
145 |
+
- ru
|
146 |
+
- sg
|
147 |
+
- sa
|
148 |
+
- sat
|
149 |
+
- scn
|
150 |
+
- shn
|
151 |
+
- si
|
152 |
+
- sk
|
153 |
+
- sl
|
154 |
+
- sm
|
155 |
+
- sn
|
156 |
+
- sd
|
157 |
+
- so
|
158 |
+
- st
|
159 |
+
- es
|
160 |
+
- sc
|
161 |
+
- sr
|
162 |
+
- ss
|
163 |
+
- su
|
164 |
+
- sv
|
165 |
+
- swh
|
166 |
+
- szl
|
167 |
+
- ta
|
168 |
+
- taq
|
169 |
+
- tt
|
170 |
+
- te
|
171 |
+
- tg
|
172 |
+
- tl
|
173 |
+
- th
|
174 |
+
- ti
|
175 |
+
- tpi
|
176 |
+
- tn
|
177 |
+
- ts
|
178 |
+
- tk
|
179 |
+
- tum
|
180 |
+
- tr
|
181 |
+
- tw
|
182 |
+
- tzm
|
183 |
+
- ug
|
184 |
+
- uk
|
185 |
+
- umb
|
186 |
+
- ur
|
187 |
+
- uzn
|
188 |
+
- vec
|
189 |
+
- vi
|
190 |
+
- war
|
191 |
+
- wo
|
192 |
+
- xh
|
193 |
+
- ydd
|
194 |
+
- yo
|
195 |
+
- yue
|
196 |
+
- zh
|
197 |
+
- zsm
|
198 |
+
- zu
|
199 |
+
|
200 |
+
language_details: "ace_Arab, ace_Latn, acm_Arab, acq_Arab, aeb_Arab, afr_Latn, ajp_Arab, aka_Latn, amh_Ethi, apc_Arab, arb_Arab, ars_Arab, ary_Arab, arz_Arab, asm_Beng, ast_Latn, awa_Deva, ayr_Latn, azb_Arab, azj_Latn, bak_Cyrl, bam_Latn, ban_Latn,bel_Cyrl, bem_Latn, ben_Beng, bho_Deva, bjn_Arab, bjn_Latn, bod_Tibt, bos_Latn, bug_Latn, bul_Cyrl, cat_Latn, ceb_Latn, ces_Latn, cjk_Latn, ckb_Arab, crh_Latn, cym_Latn, dan_Latn, deu_Latn, dik_Latn, dyu_Latn, dzo_Tibt, ell_Grek, eng_Latn, epo_Latn, est_Latn, eus_Latn, ewe_Latn, fao_Latn, pes_Arab, fij_Latn, fin_Latn, fon_Latn, fra_Latn, fur_Latn, fuv_Latn, gla_Latn, gle_Latn, glg_Latn, grn_Latn, guj_Gujr, hat_Latn, hau_Latn, heb_Hebr, hin_Deva, hne_Deva, hrv_Latn, hun_Latn, hye_Armn, ibo_Latn, ilo_Latn, ind_Latn, isl_Latn, ita_Latn, jav_Latn, jpn_Jpan, kab_Latn, kac_Latn, kam_Latn, kan_Knda, kas_Arab, kas_Deva, kat_Geor, knc_Arab, knc_Latn, kaz_Cyrl, kbp_Latn, kea_Latn, khm_Khmr, kik_Latn, kin_Latn, kir_Cyrl, kmb_Latn, kon_Latn, kor_Hang, kmr_Latn, lao_Laoo, lvs_Latn, lij_Latn, lim_Latn, lin_Latn, lit_Latn, lmo_Latn, ltg_Latn, ltz_Latn, lua_Latn, lug_Latn, luo_Latn, lus_Latn, mag_Deva, mai_Deva, mal_Mlym, mar_Deva, min_Latn, mkd_Cyrl, plt_Latn, mlt_Latn, mni_Beng, khk_Cyrl, mos_Latn, mri_Latn, zsm_Latn, mya_Mymr, nld_Latn, nno_Latn, nob_Latn, npi_Deva, nso_Latn, nus_Latn, nya_Latn, oci_Latn, gaz_Latn, ory_Orya, pag_Latn, pan_Guru, pap_Latn, pol_Latn, por_Latn, prs_Arab, pbt_Arab, quy_Latn, ron_Latn, run_Latn, rus_Cyrl, sag_Latn, san_Deva, sat_Beng, scn_Latn, shn_Mymr, sin_Sinh, slk_Latn, slv_Latn, smo_Latn, sna_Latn, snd_Arab, som_Latn, sot_Latn, spa_Latn, als_Latn, srd_Latn, srp_Cyrl, ssw_Latn, sun_Latn, swe_Latn, swh_Latn, szl_Latn, tam_Taml, tat_Cyrl, tel_Telu, tgk_Cyrl, tgl_Latn, tha_Thai, tir_Ethi, taq_Latn, taq_Tfng, tpi_Latn, tsn_Latn, tso_Latn, tuk_Latn, tum_Latn, tur_Latn, twi_Latn, tzm_Tfng, uig_Arab, ukr_Cyrl, umb_Latn, urd_Arab, uzn_Latn, vec_Latn, vie_Latn, war_Latn, wol_Latn, xho_Latn, ydd_Hebr, yor_Latn, yue_Hant, zho_Hans, zho_Hant, zul_Latn"
|
201 |
+
|
202 |
tags:
|
203 |
+
- nllb
|
204 |
+
- translation
|
205 |
+
license: "cc-by-nc-4.0"
|
206 |
+
datasets:
|
207 |
+
- flores-200
|
208 |
+
metrics:
|
209 |
+
- bleu
|
210 |
+
- spbleu
|
211 |
+
- chrf++
|
212 |
+
inference: false
|
213 |
---
|
214 |
|
215 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
216 |
should probably proofread and complete it, then remove this comment. -->
|
217 |
|
218 |
+
This model is a fine-tuned version of [facebook/nllb-200-3.3B](https://huggingface.co/facebook/nllb-200-3.3B) on the [EasyProject](https://github.com/edchengg/easyproject) dataset.
|
|
|
|
|
|
|
|
|
219 |
|
220 |
+
### Framework versions
|
|
|
|
|
|
|
|
|
221 |
|
222 |
+
- Transformers 4.29.2
|
223 |
+
- Pytorch 1.11.0+cu113
|
224 |
+
- Datasets 2.8.0
|
225 |
+
- Tokenizers 0.13.2
|
226 |
|
227 |
+
- Paper link: [Frustratingly Easy Label Projection for Cross-lingual Transfer](https://arxiv.org/abs/2211.15613)
|
228 |
+
- Github link: https://github.com/edchengg/easyproject
|
229 |
+
- Please use the transformers==4.29.2 library as Huggingface recently fixed a bug in [NLLB tokenizer](https://github.com/huggingface/transformers/pull/22313)
|
230 |
|
231 |
+
# Code
|
232 |
+
```python
|
233 |
+
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
|
234 |
+
import torch
|
235 |
|
236 |
+
tokenizer = AutoTokenizer.from_pretrained(
|
237 |
+
"facebook/nllb-200-distilled-600M", src_lang="eng_Latn")
|
238 |
+
|
239 |
+
print("Loading model")
|
240 |
+
model = AutoModelForSeq2SeqLM.from_pretrained("ychenNLP/nllb-200-3.3b-easyproject")
|
241 |
+
model.cuda()
|
242 |
|
243 |
+
input_chunks = ["A translator always risks inadvertently introducing source-language words, grammar, or syntax into the target-language rendering."]
|
244 |
+
print("Start translation...")
|
245 |
+
output_result = []
|
|
|
|
|
|
|
|
|
|
|
|
|
246 |
|
247 |
+
batch_size = 1
|
248 |
+
for idx in tqdm(range(0, len(input_chunks), batch_size)):
|
249 |
+
start_idx = idx
|
250 |
+
end_idx = idx + batch_size
|
251 |
+
inputs = tokenizer(input_chunks[start_idx: end_idx], padding=True, truncation=True, max_length=128, return_tensors="pt").to('cuda')
|
252 |
|
253 |
+
with torch.no_grad():
|
254 |
+
translated_tokens = model.generate(**inputs, forced_bos_token_id=tokenizer.lang_code_to_id["zho_Hans"],
|
255 |
+
max_length=128, num_beams=5, num_return_sequences=1, early_stopping=True)
|
256 |
|
257 |
+
output = tokenizer.batch_decode(translated_tokens, skip_special_tokens=True)
|
258 |
+
output_result.extend(output)
|
259 |
+
print(output_result)
|
260 |
+
```
|
261 |
|
262 |
+
## Citation
|
263 |
|
264 |
+
```
|
265 |
+
@inproceedings{chen2023easyproject,
|
266 |
+
title={Frustratingly Easy Label Projection for Cross-lingual Transfer},
|
267 |
+
author={Chen, Yang and Jiang, Chao and Ritter, Alan and Xu, Wei},
|
268 |
+
booktitle={Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Findings)},
|
269 |
+
year={2023}
|
270 |
+
}
|
271 |
+
```
|