jianqing666
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -2,21 +2,19 @@
|
|
2 |
license: apache-2.0
|
3 |
language:
|
4 |
- ar
|
5 |
-
- zh
|
6 |
-
- en
|
7 |
---
|
8 |
|
9 |
-
# <b>
|
10 |
MgGPT-70B-Chat is a fully fine-tuned generative text model collection, particularly focused on the Arabic language domain.
|
11 |
-
This is the repository for the 70B-chat model, developed based on [
|
12 |
|
13 |
---
|
14 |
## Model Details
|
15 |
-
We have released the
|
16 |
<!-- ## Model Developers -->
|
17 |
<!-- We are from the King Abdullah University of Science and Technology (KAUST), the Chinese University of Hong Kong, Shenzhen (CUHKSZ) and the Shenzhen Research Institute of Big Data (SRIBD). -->
|
18 |
## Variations
|
19 |
-
|
20 |
<!-- ## Paper -->
|
21 |
<!-- The paper can be accessed at [link](https://huggingface.co/FreedomIntelligence/AceGPT-v2-70B-Chat/blob/main/Alignment_at_Pre_training__a_Case_Study_of_Aligning_LLMs_in_Arabic.pdf). -->
|
22 |
## Input
|
@@ -27,11 +25,11 @@ Models output text only.
|
|
27 |
|
28 |
| Model | Avg. | [ArabicMMLU]((https://github.com/mbzuai-nlp/ArabicMMLU)) | [ArabicMMLU]((https://github.com/mbzuai-nlp/ArabicMMLU)) | ARC | BoolQ | EXAMs | ACVA (clean) | ACVA (all) |
|
29 |
|------------------|--------|--------|--------|-------|--------|--------|--------------|------------|
|
30 |
-
|
|
31 |
-
|
|
32 |
-
|
|
33 |
-
|
|
34 |
-
| **
|
35 |
| Jais-30B-v3-chat | 57.84 | 35.68 | 62.36 | 51.02 | 76.30 | 32.24 | 73.63 | 73.66 |
|
36 |
| GPT-3.5 | 62.44 | 46.07 | 57.72 | 60.24 | 76.12 | 45.63 | 74.45 | 76.88 |
|
37 |
| GPT-4 | 75.78 | 65.04 | 72.50 | 85.67 | 85.99 | 57.76 | 84.06 | 79.43 |
|
|
|
2 |
license: apache-2.0
|
3 |
language:
|
4 |
- ar
|
|
|
|
|
5 |
---
|
6 |
|
7 |
+
# <b>MgGPT0-70B-Chat</b>
|
8 |
MgGPT-70B-Chat is a fully fine-tuned generative text model collection, particularly focused on the Arabic language domain.
|
9 |
+
This is the repository for the 70B-chat model, developed based on [MgGPT0-70B](https://huggingface.co/MgGPT/MgGPT0-70B).
|
10 |
|
11 |
---
|
12 |
## Model Details
|
13 |
+
We have released the MgGPT0 family of large language models, which is a collection of fully fine-tuned generative text models based on LlaMA2(MgGPT0-7B, MgGPT0-13B), LlaMA3(MgGPT0-8B, MgGPT0-70B), Qwen2(MgGPT0-32B). Our models include two main categories: MgGPT0 and MgGPT0-chat. MgGPT0-chat is an optimized version specifically designed for dialogue applications. It is worth mentioning that our models have demonstrated superior performance compared to all currently available open-source Arabic dialogue models in multiple benchmark tests. Furthermore, in our human evaluations, our models have shown comparable satisfaction levels to some closed-source models, such as ChatGPT, in the Arabic language.
|
14 |
<!-- ## Model Developers -->
|
15 |
<!-- We are from the King Abdullah University of Science and Technology (KAUST), the Chinese University of Hong Kong, Shenzhen (CUHKSZ) and the Shenzhen Research Institute of Big Data (SRIBD). -->
|
16 |
## Variations
|
17 |
+
MgGPT0 families come in a range of parameter sizes —— 7B, 8B, 13B, 32B and 70B, each size of model has a base category and a -chat category.
|
18 |
<!-- ## Paper -->
|
19 |
<!-- The paper can be accessed at [link](https://huggingface.co/FreedomIntelligence/AceGPT-v2-70B-Chat/blob/main/Alignment_at_Pre_training__a_Case_Study_of_Aligning_LLMs_in_Arabic.pdf). -->
|
20 |
## Input
|
|
|
25 |
|
26 |
| Model | Avg. | [ArabicMMLU]((https://github.com/mbzuai-nlp/ArabicMMLU)) | [ArabicMMLU]((https://github.com/mbzuai-nlp/ArabicMMLU)) | ARC | BoolQ | EXAMs | ACVA (clean) | ACVA (all) |
|
27 |
|------------------|--------|--------|--------|-------|--------|--------|--------------|------------|
|
28 |
+
| MgGPT0-7B-chat | 59.90 | 45.77 | 56.62 | 60.49 | 72.45 | 43.69 | 69.46 | 70.86 |
|
29 |
+
| MgGPT0-13B-chat | 63.42 | 47.33 | 61.70 | 63.99 | 69.33 | 48.37 | 76.90 | 76.37 |
|
30 |
+
| MgGPT0-8B-chat | 66.69 | 54.45 | 62.21 | 72.44 | 71.65 | 52.98 | 76.54 | 76.55 |
|
31 |
+
| MgGPT0-32B-chat | 70.63 | 57.12 | 68.70 | 78.07 | 77.22 | 52.89 | 81.36 | 79.03 |
|
32 |
+
| **MgGPT0-70B-chat** | 73.99 | 64.26 | 72.50 | 85.53 | 82.66 | 56.99 | 78.61 | 77.38 |
|
33 |
| Jais-30B-v3-chat | 57.84 | 35.68 | 62.36 | 51.02 | 76.30 | 32.24 | 73.63 | 73.66 |
|
34 |
| GPT-3.5 | 62.44 | 46.07 | 57.72 | 60.24 | 76.12 | 45.63 | 74.45 | 76.88 |
|
35 |
| GPT-4 | 75.78 | 65.04 | 72.50 | 85.67 | 85.99 | 57.76 | 84.06 | 79.43 |
|