Fairseq
Italian
Catalan
AudreyVM commited on
Commit
5f5f938
1 Parent(s): bafedc7

update model-card

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -48,7 +48,7 @@ import pyonmttok
48
  from huggingface_hub import snapshot_download
49
  model_dir = snapshot_download(repo_id="projecte-aina/mt-aina-it-ca", revision="main")
50
  tokenizer=pyonmttok.Tokenizer(mode="none", sp_model_path = model_dir + "/spm.model")
51
- tokenized=tokenizer.tokenize("Benvingut al projecte Aina!")
52
  translator = ctranslate2.Translator(model_dir)
53
  translated = translator.translate_batch([tokenized[0]])
54
  print(tokenizer.detokenize(translated[0][0]['tokens']))
@@ -117,9 +117,9 @@ We use the BLEU score for evaluation on the Flores test set: [Flores-101](https:
117
  Below are the evaluation results on the machine translation from Catalan to Italian compared to [Softcatalà](https://www.softcatala.org/) and [Google Translate](https://translate.google.es/?hl=es):
118
  | Test set | SoftCatalà | Google Translate |mt-aina-it-ca|
119
  |----------------------|------------|------------------|---------------|
120
- | Flores 101 dev | 25,4 | **30,4** | 27,4 |
121
- | Flores 101 devtest |26,6 | **31,2** | 27,9 |
122
- | Average | 26,0 | **30,8** | 27,7 |
123
  ## Additional information
124
  ### Author
125
  Language Technologies Unit (LangTech) at the Barcelona Supercomputing Center (langtech@bsc.es)
 
48
  from huggingface_hub import snapshot_download
49
  model_dir = snapshot_download(repo_id="projecte-aina/mt-aina-it-ca", revision="main")
50
  tokenizer=pyonmttok.Tokenizer(mode="none", sp_model_path = model_dir + "/spm.model")
51
+ tokenized=tokenizer.tokenize("Benvenuto al progetto Aina!")
52
  translator = ctranslate2.Translator(model_dir)
53
  translated = translator.translate_batch([tokenized[0]])
54
  print(tokenizer.detokenize(translated[0][0]['tokens']))
 
117
  Below are the evaluation results on the machine translation from Catalan to Italian compared to [Softcatalà](https://www.softcatala.org/) and [Google Translate](https://translate.google.es/?hl=es):
118
  | Test set | SoftCatalà | Google Translate |mt-aina-it-ca|
119
  |----------------------|------------|------------------|---------------|
120
+ | Flores 101 dev | 25,4 | **30,4** | 26,6 |
121
+ | Flores 101 devtest |26,6 | **31,2** | 27,2 |
122
+ | Average | 26,0 | **30,8** | 29,6 |
123
  ## Additional information
124
  ### Author
125
  Language Technologies Unit (LangTech) at the Barcelona Supercomputing Center (langtech@bsc.es)