Upload README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,73 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
2 |
license: cc-by-nc-sa-4.0
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
language:
|
3 |
+
- ko
|
4 |
+
library_name: transformers
|
5 |
+
pipeline_tag: text-generation
|
6 |
license: cc-by-nc-sa-4.0
|
7 |
+
datasets:
|
8 |
+
- kyujinpy/KOR-OpenOrca-Platypus-v3
|
9 |
---
|
10 |
+
|
11 |
+
# **PracticeLLM/KoSOLAR-Platypus-10.7B**
|
12 |
+
|
13 |
+
## Model Details
|
14 |
+
|
15 |
+
**Model Developers** Kyujin Han (kyujinpy)
|
16 |
+
|
17 |
+
**Method**
|
18 |
+
LoRA with quantization.
|
19 |
+
|
20 |
+
**Dataset**
|
21 |
+
[kyujinpy/KOR-OpenOrca-Platypus-v3](https://huggingface.co/datasets/kyujinpy/KOR-OpenOrca-Platypus-v3).
|
22 |
+
|
23 |
+
**Hyperparameters**
|
24 |
+
```
|
25 |
+
python finetune.py \
|
26 |
+
--base_model yanolja/KoSOLAR-10.7B-v0.2 \
|
27 |
+
--data-path kyujinpy/KOR-OpenOrca-Platypus-v3 \
|
28 |
+
--output_dir ./Ko-PlatypusSOLAR-10.7B \
|
29 |
+
--batch_size 64 \
|
30 |
+
--micro_batch_size 1 \
|
31 |
+
--num_epochs 5 \
|
32 |
+
--learning_rate 2e-5 \
|
33 |
+
--cutoff_len 2048 \
|
34 |
+
--val_set_size 0 \
|
35 |
+
--lora_r 64 \
|
36 |
+
--lora_alpha 64 \
|
37 |
+
--lora_dropout 0.05 \
|
38 |
+
--lora_target_modules '[embed_tokens, q_proj, k_proj, v_proj, o_proj, gate_proj, down_proj, up_proj, lm_head]' \
|
39 |
+
--train_on_inputs False \
|
40 |
+
--add_eos_token False \
|
41 |
+
--group_by_length False \
|
42 |
+
--prompt_template_name en_simple \
|
43 |
+
--lr_scheduler 'cosine' \
|
44 |
+
```
|
45 |
+
> Share all of things. It is my belief.
|
46 |
+
|
47 |
+
# **Model Benchmark**
|
48 |
+
|
49 |
+
## Open Ko-LLM leaderboard & lm-evaluation-harness(zero-shot)
|
50 |
+
- Follow up as [Ko-link](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard).
|
51 |
+
| Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Ko-CommonGenV2 |
|
52 |
+
| --- | --- | --- | --- | --- | --- | --- |
|
53 |
+
| PracticeLLM/KoSOLAR-Platypus-10.7B | --- | --- | --- | --- | --- | --- |
|
54 |
+
| [LDCC/LDCC-SOLAR-10.7B](https://huggingface.co/LDCC/LDCC-SOLAR-10.7B) | 59.34 | 55.38 | 65.56 | 53.38 | 64.39 | 57.97 |
|
55 |
+
| [yanolja/KoSOLAR-10.7B-v0.2](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.2) | 55.62 | 50.51 | 62.29 | 53.76 | 47.31 | 64.23 |
|
56 |
+
| [megastudyedu/M-SOLAR-10.7B-v1.3](https://huggingface.co/megastudyedu/M-SOLAR-10.7B-v1.3) | 56.64 | 51.37 | 60.93 | 54.91 | 48.45 | 67.53 |
|
57 |
+
|
58 |
+
|
59 |
+
# Implementation Code
|
60 |
+
```python
|
61 |
+
### KO-Platypus
|
62 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
63 |
+
import torch
|
64 |
+
|
65 |
+
repo = "PracticeLLM/KoSOLAR-Platypus-10.7B"
|
66 |
+
OpenOrca = AutoModelForCausalLM.from_pretrained(
|
67 |
+
repo,
|
68 |
+
return_dict=True,
|
69 |
+
torch_dtype=torch.float16,
|
70 |
+
device_map='auto'
|
71 |
+
)
|
72 |
+
OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo)
|
73 |
+
```
|