nonetrix commited on
Commit
b06d31c
1 Parent(s): b8fcbee

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -5
README.md CHANGED
@@ -1,17 +1,28 @@
1
  ---
2
  base_model:
3
- - SanjiWatsuki/Kunoichi-DPO-v2-7B
4
- - sanjiwatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE
5
  - Endevor/EndlessRP-v3-7B
6
- - yam-peleg/Experiment30-7B
 
7
  - undi95/Toppy-M-7B
 
8
  library_name: transformers
9
  tags:
10
  - mergekit
11
  - merge
12
-
 
13
  ---
14
- # merge
 
 
 
 
 
 
 
 
 
 
15
 
16
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
17
 
 
1
  ---
2
  base_model:
 
 
3
  - Endevor/EndlessRP-v3-7B
4
+ - sanjiwatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE
5
+ - SanjiWatsuki/Kunoichi-DPO-v2-7B
6
  - undi95/Toppy-M-7B
7
+ - yam-peleg/Experiment30-7B
8
  library_name: transformers
9
  tags:
10
  - mergekit
11
  - merge
12
+ - not-for-all-audiences
13
+ license: apache-2.0
14
  ---
15
+ # Pippafeet-11B-0.2
16
+ This model is a mix of some of the "best 7B roleplaying LLMs". I selected a few models based on "creativity" from a random benchmark, and a final roleplaying LLM based on "IQ," and finally another LLM merged twice that "excels at general tasks" according to a separate benchmark for it's size. My goal was to combine the "most creative" smaller roleplaying LLMs, merge them, and enhance the intelligence by incorporating "decent general model" twice, along with a "smarter" roleplaying LLM. I don't really trust benchmarks much, but I thought it would at least give it some alignment perhaps, even if it is overfitted to a dataset to score well, I thought since it's a merge so it might negate overfitting somewhat, seems to have worked to some extent luckily. Also, this is a iteration of [my previous model](https://huggingface.co/nonetrix/pippafeet-11B-0.1) with only thing changed being the merge method. This iteration seems to be somewhat of a sidegrade rather than a pure upgrade, but for the most part it is more stable and generally better.
17
+
18
+ In my limited testing, this model performs really well, giving decent replies most of the time.... That is if you ignore the fatal flaws, which are inherent to how this model was created unfortunately. Since it's made by stacking the weights of other models, it likes to constantly create new words and stutter and generally act stange, however if you ignore this and fill in the blanks yourself the model is quite decent. I plan to try to remove this weirdness if possible by using a LoRA but I am not sure if I will be able to, no promisses. If you have the compute to fine tune this model, I emplore you to because I think it is a promissing base.
19
+
20
+ Edit: Fine tune is pretty much impossible because ROCm is hot garbage and I should have never bought a AMD GPU, if someone has a functional GPU please fine tune it for me. Might be able to do it on CPU somehow but likely not FP16 and slow as fuck and in GGUF
21
+
22
+ Artwork source, please contact for me to remove it if wished: https://twitter.com/Kumaartsu/status/1756793643384402070
23
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6342619a9948f573f37a4a60/MUD4N762ncyUw2dPfzVJ_.png)
24
+
25
+ Note: this model is in no way affiliated with Phase Connect, Pipkin Pippa, or the artists artwork.
26
 
27
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
28