File size: 1,250 Bytes
6ff644c
57e6c38
6ff644c
 
57e6c38
6ff644c
8cb44f1
6ff644c
57e6c38
6ff644c
f6b6c49
 
 
 
 
 
 
 
 
 
69904ce
f6b6c49
 
57e6c38
 
 
c9dba8d
 
57e6c38
 
 
 
 
 
 
 
86f2e9c
57e6c38
 
 
 
 
 
 
 
 
8ddd071
52c1725
 
8ddd071
 
 
 
 
57e6c38
 
 
a9d285d
 
57e6c38
86f2e9c
57e6c38
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
---
license: "apache-2.0"
---

*This model was trained as part of a series of experiments testing the performance of pure DPO vs SFT vs ORPO, all supported by Unsloth/Huggingface TRL.*

Note: Extremely buggy, not recommended for use.

**Benchmarks**

Average 59.63

ARC 59.47

HellaSwag 82.47

MMLU 62.31

TruthfulQA 40.11

Winogrande 78.3

GSM8K 35.1

**Training Details**

Duration: ~10-12 hours on one Kaggle T4 with Unsloth

Model: https://huggingface.co./unsloth/mistral-7b-v0.2-bnb-4bit

Dataset: https://huggingface.co./datasets/argilla/dpo-mix-7k 

Rank: 8 

Alpha: 16 

Learning rate: 5e-5

Beta: 0.1 

Batch size: 8 

Epochs: 1

Learning rate scheduler: Linear

Prompt Format: ChatML
```
<|im_start|>system
You are a helpful assistant.<|im_end|>
<|im_start|>user
Why is the sky blue?<|im_end|>
<|im_start|>assistant
```


**WanDB Reports**

![image/png](https://cdn-uploads.huggingface.co/production/uploads/65a5c0e82823ba72ed2cee7d/CIoGd-61duUAKcnlaAGR6.png)

![image/png](https://cdn-uploads.huggingface.co/production/uploads/65a5c0e82823ba72ed2cee7d/5ks-ZGg1jFNp821P2Si3S.png)

[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)