Update README.md
Browse files
README.md
CHANGED
@@ -11,10 +11,9 @@ license: cc-by-nc-4.0
|
|
11 |
# kuno-kunoichi-v1-DPO-v2-SLERP-7B
|
12 |
|
13 |
kuno-kunoichi-v1-DPO-v2-SLERP-7B is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
14 |
-
I'm hoping that the result is more robust against errors, as the two models likely implement comparable reasoning at least somewhat differently.
|
15 |
-
|
16 |
-
I've performed some testing with ChatML format prompting using temperature=1.1 and minP=0.03.
|
17 |
|
|
|
18 |
## Merge Details
|
19 |
### Merge Method
|
20 |
|
|
|
11 |
# kuno-kunoichi-v1-DPO-v2-SLERP-7B
|
12 |
|
13 |
kuno-kunoichi-v1-DPO-v2-SLERP-7B is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
14 |
+
I'm hoping that the result is more robust against errors or when merging due to "denseness", as the two models likely implement comparable reasoning at least somewhat differently.
|
|
|
|
|
15 |
|
16 |
+
I've performed some testing with ChatML format prompting using temperature=1.1 and minP=0.03. The model also supports Alpaca format promtps.
|
17 |
## Merge Details
|
18 |
### Merge Method
|
19 |
|