Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Sean Cho
commited on
Commit
•
59ba8a1
1
Parent(s):
c770c05
update text
Browse files
src/assets/text_content.py
CHANGED
@@ -32,7 +32,9 @@ Please provide information about the model through an issue! 🤩
|
|
32 |
|
33 |
## How it works
|
34 |
|
35 |
-
📈 We
|
|
|
|
|
36 |
- Ko-HellaSwag (provided by __[Upstage](https://www.upstage.ai/)__, machine translation)
|
37 |
- Ko-MMLU (provided by __[Upstage](https://www.upstage.ai/)__, human translation and variation)
|
38 |
- Ko-Arc (provided by __[Upstage](https://www.upstage.ai/)__, human translation and variation)
|
@@ -66,7 +68,11 @@ tokenizer = AutoTokenizer.from_pretrained("your model name", revision=revision)
|
|
66 |
```
|
67 |
|
68 |
If this step fails, follow the error messages to debug your model before submitting it. It's likely your model has been improperly uploaded.
|
|
|
69 |
⚠️ Make sure your model is public!
|
|
|
|
|
|
|
70 |
⚠️ If your model needs use_remote_code=True, we do not support this option yet but we are working on adding it, stay posted!
|
71 |
|
72 |
### 2️⃣ Convert your model weights to [safetensors](https://huggingface.co/docs/safetensors/index)
|
|
|
32 |
|
33 |
## How it works
|
34 |
|
35 |
+
📈 We evaluate models using the [Eleuther AI Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness), a unified framework to test generative language models on a large number of different evaluation tasks.
|
36 |
+
|
37 |
+
We have set up a benchmark using datasets translated into Korean, and applied variations by human experts, from the four tasks (HellaSwag, MMLU, Arc, Truthful QA) operated by HuggingFace OpenLLM. We have also added a new dataset prepared from scratch.
|
38 |
- Ko-HellaSwag (provided by __[Upstage](https://www.upstage.ai/)__, machine translation)
|
39 |
- Ko-MMLU (provided by __[Upstage](https://www.upstage.ai/)__, human translation and variation)
|
40 |
- Ko-Arc (provided by __[Upstage](https://www.upstage.ai/)__, human translation and variation)
|
|
|
68 |
```
|
69 |
|
70 |
If this step fails, follow the error messages to debug your model before submitting it. It's likely your model has been improperly uploaded.
|
71 |
+
|
72 |
⚠️ Make sure your model is public!
|
73 |
+
|
74 |
+
⚠️ Maker sure your model runs with [Eleuther AI Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness)
|
75 |
+
|
76 |
⚠️ If your model needs use_remote_code=True, we do not support this option yet but we are working on adding it, stay posted!
|
77 |
|
78 |
### 2️⃣ Convert your model weights to [safetensors](https://huggingface.co/docs/safetensors/index)
|