chaoweihuang commited on
Commit
3179552
1 Parent(s): f097e90

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -13
README.md CHANGED
@@ -15,7 +15,24 @@ model-index:
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
  should probably proofread and complete it, then remove this comment. -->
17
 
18
- # kto-mix-14k-lf-response-phi3-f1_100_0.7-fg0.5-kto-fg-fgudw4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on the trl-lib/kto-mix-14k and the chaoweihuang/lf-response-phi3-f1_100_0.7-fg0.5 datasets.
21
  It achieves the following results on the evaluation set:
@@ -39,18 +56,6 @@ It achieves the following results on the evaluation set:
39
  - Fg Logps/reference Kl: -20.2070
40
  - Fg Loss: 0.7365
41
 
42
- ## Model description
43
-
44
- More information needed
45
-
46
- ## Intended uses & limitations
47
-
48
- More information needed
49
-
50
- ## Training and evaluation data
51
-
52
- More information needed
53
-
54
  ## Training procedure
55
 
56
  ### Training hyperparameters
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
  should probably proofread and complete it, then remove this comment. -->
17
 
18
+ # FactAlign-Phi-3-Mini
19
+
20
+ This model is aligned with our **FactAlign** framework for improved long-form factuality, from [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct).
21
+
22
+ For more information, please refer to our paper: [FactAlign: Long-form Factuality Alignment of Large Language Models](https://huggingface.co/papers/2410.01691).
23
+
24
+
25
+
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
 
37
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on the trl-lib/kto-mix-14k and the chaoweihuang/lf-response-phi3-f1_100_0.7-fg0.5 datasets.
38
  It achieves the following results on the evaluation set:
 
56
  - Fg Logps/reference Kl: -20.2070
57
  - Fg Loss: 0.7365
58
 
 
 
 
 
 
 
 
 
 
 
 
 
59
  ## Training procedure
60
 
61
  ### Training hyperparameters