elmadany commited on
Commit
6c7050f
·
1 Parent(s): 7da73b8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -3
README.md CHANGED
@@ -1,13 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  # AraT5v2-base-1024
2
 
3
  ## What's new?
4
- - **More data.** AraT2v2 trained on multiple varieties of Arabic data.
5
  - **Large sequence length.** We increase the sequence length from 512 to 1024 in this version.
6
- - **Converge faster.** AraT5v2 converges more than 10x compared with the previous version (AraT5-base.
7
- - **Extra IDs.** AraT5v2 supports 100 sentinel tokens (a.k.a unique mask tokens).
8
 
9
  ```diff
10
  - We recommend using this version (AraT5v2-base-1024) instead of the previous version (AraT5-base).
11
  ```
12
 
13
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ar
4
+ tags:
5
+ - Arabic T5
6
+ - MSA
7
+ - Twitter
8
+ - Arabic Dialect
9
+ - Arabic Machine Translation
10
+ - Arabic Text Summarization
11
+ - Arabic News Title and Question Generation
12
+ - Arabic Paraphrasing and Transliteration
13
+ - Arabic Code-Switched Translation
14
+ ---
15
  # AraT5v2-base-1024
16
 
17
  ## What's new?
18
+ - **More data.** AraT5v2-base-1024 trained on multiple varieties of Arabic data.
19
  - **Large sequence length.** We increase the sequence length from 512 to 1024 in this version.
20
+ - **Converge faster.** AraT5v2-base-1024 converges more than 10x compared with the previous version (AraT5-base.
21
+ - **Extra IDs.** AraT5v2-base-1024 supports 100 sentinel tokens (a.k.a unique mask tokens).
22
 
23
  ```diff
24
  - We recommend using this version (AraT5v2-base-1024) instead of the previous version (AraT5-base).
25
  ```
26
 
27
 
28
+
29
+
30
+
31
+ # Citation
32
+
33
+ If you use our models (AraT5v2-base-1024, Arat5-base, Arat5-msa-base, Arat5-tweet-base, Arat5-msa-small, or Arat5-tweet-small ) for your scientific publication, or if you find the resources in this repository useful, please cite our paper as follows (to be updated):
34
+ ```bibtex
35
+ @inproceedings{nagoudi2022_arat5,
36
+ @inproceedings{nagoudi-etal-2022-arat5,
37
+ title = "{A}ra{T}5: Text-to-Text Transformers for {A}rabic Language Generation",
38
+ author = "Nagoudi, El Moatez Billah and
39
+ Elmadany, AbdelRahim and
40
+ Abdul-Mageed, Muhammad",
41
+ booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
42
+ month = may,
43
+ year = "2022",
44
+ address = "Dublin, Ireland",
45
+ publisher = "Association for Computational Linguistics",
46
+ url = "https://aclanthology.org/2022.acl-long.47",
47
+ pages = "628--647",
48
+ abstract = "Transfer learning with a unified Transformer framework (T5) that converts all language problems into a text-to-text format was recently proposed as a simple and effective transfer learning approach. Although a multilingual version of the T5 model (mT5) was also introduced, it is not clear how well it can fare on non-English tasks involving diverse data. To investigate this question, we apply mT5 on a language with a wide variety of dialects{--}Arabic. For evaluation, we introduce a novel benchmark for ARabic language GENeration (ARGEN), covering seven important tasks. For model comparison, we pre-train three powerful Arabic T5-style models and evaluate them on ARGEN. Although pre-trained with {\textasciitilde}49 less data, our new models perform significantly better than mT5 on all ARGEN tasks (in 52 out of 59 test sets) and set several new SOTAs. Our models also establish new SOTA on the recently-proposed, large Arabic language understanding evaluation benchmark ARLUE (Abdul-Mageed et al., 2021). Our new models are publicly available. We also link to ARGEN datasets through our repository: https://github.com/UBC-NLP/araT5.",
49
+ }
50
+