elmadany commited on
Commit
b99444f
1 Parent(s): 62e7862

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -2
README.md CHANGED
@@ -62,7 +62,9 @@ AraT5 Pytorch and TensorFlow checkpoints are available on the Huggingface websit
62
 
63
  # BibTex
64
 
65
- If you use our models (Arat5-base, Arat5-msa-base, Arat5-tweet-base, Arat5-msa-small, or Arat5-tweet-small ) for your scientific publication, or if you find the resources in this repository useful, please cite our paper as follows (to be updated):
 
 
66
  ```bibtex
67
  @inproceedings{nagoudi2022_arat5,
68
  @inproceedings{nagoudi-etal-2022-arat5,
@@ -79,6 +81,23 @@ If you use our models (Arat5-base, Arat5-msa-base, Arat5-tweet-base, Arat5-msa-s
79
  pages = "628--647",
80
  abstract = "Transfer learning with a unified Transformer framework (T5) that converts all language problems into a text-to-text format was recently proposed as a simple and effective transfer learning approach. Although a multilingual version of the T5 model (mT5) was also introduced, it is not clear how well it can fare on non-English tasks involving diverse data. To investigate this question, we apply mT5 on a language with a wide variety of dialects{--}Arabic. For evaluation, we introduce a novel benchmark for ARabic language GENeration (ARGEN), covering seven important tasks. For model comparison, we pre-train three powerful Arabic T5-style models and evaluate them on ARGEN. Although pre-trained with {\textasciitilde}49 less data, our new models perform significantly better than mT5 on all ARGEN tasks (in 52 out of 59 test sets) and set several new SOTAs. Our models also establish new SOTA on the recently-proposed, large Arabic language understanding evaluation benchmark ARLUE (Abdul-Mageed et al., 2021). Our new models are publicly available. We also link to ARGEN datasets through our repository: https://github.com/UBC-NLP/araT5.",
81
  }
82
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
83
  ## Acknowledgments
84
  We gratefully acknowledge support from the Natural Sciences and Engineering Research Council of Canada, the Social Sciences and Humanities Research Council of Canada, Canadian Foundation for Innovation, [ComputeCanada](www.computecanada.ca) and [UBC ARC-Sockeye](https://doi.org/10.14288/SOCKEYE). We also thank the [Google TensorFlow Research Cloud (TFRC)](https://www.tensorflow.org/tfrc) program for providing us with free TPU access.
 
62
 
63
  # BibTex
64
 
65
+ If you use our models for your scientific publication, or if you find the resources in this repository useful, please cite our papers as follows:
66
+
67
+ **(AraT5-base, AraT5-msa-base, AraT5-tweet-base, AraT5-msa-small, or AraT5-tweet-small)**
68
  ```bibtex
69
  @inproceedings{nagoudi2022_arat5,
70
  @inproceedings{nagoudi-etal-2022-arat5,
 
81
  pages = "628--647",
82
  abstract = "Transfer learning with a unified Transformer framework (T5) that converts all language problems into a text-to-text format was recently proposed as a simple and effective transfer learning approach. Although a multilingual version of the T5 model (mT5) was also introduced, it is not clear how well it can fare on non-English tasks involving diverse data. To investigate this question, we apply mT5 on a language with a wide variety of dialects{--}Arabic. For evaluation, we introduce a novel benchmark for ARabic language GENeration (ARGEN), covering seven important tasks. For model comparison, we pre-train three powerful Arabic T5-style models and evaluate them on ARGEN. Although pre-trained with {\textasciitilde}49 less data, our new models perform significantly better than mT5 on all ARGEN tasks (in 52 out of 59 test sets) and set several new SOTAs. Our models also establish new SOTA on the recently-proposed, large Arabic language understanding evaluation benchmark ARLUE (Abdul-Mageed et al., 2021). Our new models are publicly available. We also link to ARGEN datasets through our repository: https://github.com/UBC-NLP/araT5.",
83
  }
84
+ ```
85
+ **AraT5v2-base-1024**
86
+ ``` bibtex
87
+ @inproceedings{elmadany-etal-2023-octopus,
88
+ title = "Octopus: A Multitask Model and Toolkit for {A}rabic Natural Language Generation",
89
+ author = "Elmadany, AbdelRahim and
90
+ Nagoudi, El Moatez Billah and
91
+ Abdul-Mageed, Muhammad",
92
+ booktitle = "Proceedings of ArabicNLP 2023",
93
+ month = dec,
94
+ year = "2023",
95
+ address = "Singapore (Hybrid)",
96
+ publisher = "Association for Computational Linguistics",
97
+ url = "https://aclanthology.org/2023.arabicnlp-1.20",
98
+ doi = "10.18653/v1/2023.arabicnlp-1.20",
99
+ pages = "232--243",
100
+ }
101
+ ```
102
  ## Acknowledgments
103
  We gratefully acknowledge support from the Natural Sciences and Engineering Research Council of Canada, the Social Sciences and Humanities Research Council of Canada, Canadian Foundation for Innovation, [ComputeCanada](www.computecanada.ca) and [UBC ARC-Sockeye](https://doi.org/10.14288/SOCKEYE). We also thank the [Google TensorFlow Research Cloud (TFRC)](https://www.tensorflow.org/tfrc) program for providing us with free TPU access.