update model-card
Browse files
README.md
CHANGED
@@ -48,7 +48,7 @@ import pyonmttok
|
|
48 |
from huggingface_hub import snapshot_download
|
49 |
model_dir = snapshot_download(repo_id="projecte-aina/mt-aina-it-ca", revision="main")
|
50 |
tokenizer=pyonmttok.Tokenizer(mode="none", sp_model_path = model_dir + "/spm.model")
|
51 |
-
tokenized=tokenizer.tokenize("
|
52 |
translator = ctranslate2.Translator(model_dir)
|
53 |
translated = translator.translate_batch([tokenized[0]])
|
54 |
print(tokenizer.detokenize(translated[0][0]['tokens']))
|
@@ -117,9 +117,9 @@ We use the BLEU score for evaluation on the Flores test set: [Flores-101](https:
|
|
117 |
Below are the evaluation results on the machine translation from Catalan to Italian compared to [Softcatalà](https://www.softcatala.org/) and [Google Translate](https://translate.google.es/?hl=es):
|
118 |
| Test set | SoftCatalà | Google Translate |mt-aina-it-ca|
|
119 |
|----------------------|------------|------------------|---------------|
|
120 |
-
| Flores 101 dev | 25,4 | **30,4** |
|
121 |
-
| Flores 101 devtest |26,6 | **31,2** | 27,
|
122 |
-
| Average | 26,0 | **30,8** |
|
123 |
## Additional information
|
124 |
### Author
|
125 |
Language Technologies Unit (LangTech) at the Barcelona Supercomputing Center (langtech@bsc.es)
|
|
|
48 |
from huggingface_hub import snapshot_download
|
49 |
model_dir = snapshot_download(repo_id="projecte-aina/mt-aina-it-ca", revision="main")
|
50 |
tokenizer=pyonmttok.Tokenizer(mode="none", sp_model_path = model_dir + "/spm.model")
|
51 |
+
tokenized=tokenizer.tokenize("Benvenuto al progetto Aina!")
|
52 |
translator = ctranslate2.Translator(model_dir)
|
53 |
translated = translator.translate_batch([tokenized[0]])
|
54 |
print(tokenizer.detokenize(translated[0][0]['tokens']))
|
|
|
117 |
Below are the evaluation results on the machine translation from Catalan to Italian compared to [Softcatalà](https://www.softcatala.org/) and [Google Translate](https://translate.google.es/?hl=es):
|
118 |
| Test set | SoftCatalà | Google Translate |mt-aina-it-ca|
|
119 |
|----------------------|------------|------------------|---------------|
|
120 |
+
| Flores 101 dev | 25,4 | **30,4** | 26,6 |
|
121 |
+
| Flores 101 devtest |26,6 | **31,2** | 27,2 |
|
122 |
+
| Average | 26,0 | **30,8** | 29,6 |
|
123 |
## Additional information
|
124 |
### Author
|
125 |
Language Technologies Unit (LangTech) at the Barcelona Supercomputing Center (langtech@bsc.es)
|