kyujinpy commited on
Commit
ea5f015
1 Parent(s): 0a09770

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -0
README.md CHANGED
@@ -1,3 +1,73 @@
1
  ---
 
 
 
 
2
  license: cc-by-nc-sa-4.0
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - ko
4
+ library_name: transformers
5
+ pipeline_tag: text-generation
6
  license: cc-by-nc-sa-4.0
7
+ datasets:
8
+ - kyujinpy/KOR-OpenOrca-Platypus-v3
9
  ---
10
+
11
+ # **PracticeLLM/KoSOLAR-Platypus-10.7B**
12
+
13
+ ## Model Details
14
+
15
+ **Model Developers** Kyujin Han (kyujinpy)
16
+
17
+ **Method**
18
+ LoRA with quantization.
19
+
20
+ **Dataset**
21
+ [kyujinpy/KOR-OpenOrca-Platypus-v3](https://huggingface.co/datasets/kyujinpy/KOR-OpenOrca-Platypus-v3).
22
+
23
+ **Hyperparameters**
24
+ ```
25
+ python finetune.py \
26
+ --base_model yanolja/KoSOLAR-10.7B-v0.2 \
27
+ --data-path kyujinpy/KOR-OpenOrca-Platypus-v3 \
28
+ --output_dir ./Ko-PlatypusSOLAR-10.7B \
29
+ --batch_size 64 \
30
+ --micro_batch_size 1 \
31
+ --num_epochs 5 \
32
+ --learning_rate 2e-5 \
33
+ --cutoff_len 2048 \
34
+ --val_set_size 0 \
35
+ --lora_r 64 \
36
+ --lora_alpha 64 \
37
+ --lora_dropout 0.05 \
38
+ --lora_target_modules '[embed_tokens, q_proj, k_proj, v_proj, o_proj, gate_proj, down_proj, up_proj, lm_head]' \
39
+ --train_on_inputs False \
40
+ --add_eos_token False \
41
+ --group_by_length False \
42
+ --prompt_template_name en_simple \
43
+ --lr_scheduler 'cosine' \
44
+ ```
45
+ > Share all of things. It is my belief.
46
+
47
+ # **Model Benchmark**
48
+
49
+ ## Open Ko-LLM leaderboard & lm-evaluation-harness(zero-shot)
50
+ - Follow up as [Ko-link](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard).
51
+ | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Ko-CommonGenV2 |
52
+ | --- | --- | --- | --- | --- | --- | --- |
53
+ | PracticeLLM/KoSOLAR-Platypus-10.7B | --- | --- | --- | --- | --- | --- |
54
+ | [LDCC/LDCC-SOLAR-10.7B](https://huggingface.co/LDCC/LDCC-SOLAR-10.7B) | 59.34 | 55.38 | 65.56 | 53.38 | 64.39 | 57.97 |
55
+ | [yanolja/KoSOLAR-10.7B-v0.2](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.2) | 55.62 | 50.51 | 62.29 | 53.76 | 47.31 | 64.23 |
56
+ | [megastudyedu/M-SOLAR-10.7B-v1.3](https://huggingface.co/megastudyedu/M-SOLAR-10.7B-v1.3) | 56.64 | 51.37 | 60.93 | 54.91 | 48.45 | 67.53 |
57
+
58
+
59
+ # Implementation Code
60
+ ```python
61
+ ### KO-Platypus
62
+ from transformers import AutoModelForCausalLM, AutoTokenizer
63
+ import torch
64
+
65
+ repo = "PracticeLLM/KoSOLAR-Platypus-10.7B"
66
+ OpenOrca = AutoModelForCausalLM.from_pretrained(
67
+ repo,
68
+ return_dict=True,
69
+ torch_dtype=torch.float16,
70
+ device_map='auto'
71
+ )
72
+ OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo)
73
+ ```