File size: 4,046 Bytes
1e675e6 bacc0ef 1e675e6 03c1898 3a26373 c98c332 1e675e6 c47c3aa 1e675e6 c47c3aa ae5398b 1e675e6 0bed6fc 1e675e6 0cdcba8 1e675e6 090c159 c31f8f2 e6a6ef1 1e675e6 c98c332 1e675e6 090c159 1e675e6 c47c3aa 1e675e6 1b53c31 1e675e6 350e392 1e675e6 8953908 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 |
---
license: cc-by-nc-4.0
---
## There's some kind of gguf tokenization issue where its not properly handling the <|im_end|> EOS token and instead breaks it into multiple smaller ones which degrades performance for gguf users. Need to submit an issue ticket. Unsure if this impacts other quantization formats.
## If possible, please use the safetensors or Exllama2 for best performance. Exllama2 tokenizes correctly.
Thank you to Lucyknada for Exl2 Quants
https://huggingface.co./lucyknada/cgato_Nemo-12b-Humanize-KTO-v0.1-exl2
[<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
This is the first public release of the Humanize model I've been working on. That way folks have a stable repo to download from and can expect some consistient level of performance when downloading.
## Goals
The goal of this model is to make something that feels different from other models out there. I wanted to accomplish this in two ways, first use as little synthetic data as possible, and second, by hand judging KTO samples to avoid common AI phrases and repetition. It has seen very very little synthetic data and without RL is unusable. The RL dataset was built by pulling responses from the model itself and then having a human (me) review each response.
The KTO dataset for this model consisted of 7122 human judged responses. Of the 7122 responses, 2398 were accepted and 4724 were rejected. To balance this, the KTO finetune was run using a 2.5x weight for desirable responses.
One result of using so much human data is that this model does not call itself ChatGPT or Claude. Examples below.
## Chat Format: ChatML with System Prompt
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64dd7cda3d6b954bf7cdd922/B9z8ABceTuaBEJAZr4hWl.png)
Logits taken for the above input with samplers disabled.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64dd7cda3d6b954bf7cdd922/Zo3IUVcr8dZi9dt4ETuZr.png)
Some sample completions taken with the Sampler settings listed at the end of the document.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64dd7cda3d6b954bf7cdd922/gJ3QXw7hOlQcGkshsM1Yt.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64dd7cda3d6b954bf7cdd922/ccH29PPsRqmoEAqqrx7ar.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64dd7cda3d6b954bf7cdd922/JjsuGBmoYCdXlOfIi1KZM.png)
Note that if you tell it that it's an AI and ask it for its name, it will call itself Claude/GPT. I did not tune to encourage or remove this behavior.
## Issues
* GGUF currently does not tokenize the end token correctly, resulting in lower performance for GGUF users.
* Male characters are not thoroughly represented in the RL dataset.
* Additional work needs to be done polishing assistant answers and performance.
* Model sometimes writes for the user during the first few responses, will probably be solved with more data. For now just regenerate.
## All samples taken at bfloat16, Transformers with FA2.
## Examples Roleplay
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64dd7cda3d6b954bf7cdd922/Ms2sV2xJUDkJKxobx7hQG.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64dd7cda3d6b954bf7cdd922/27EK2APsD6_w46EDiHNUT.png)
## Example Chat
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64dd7cda3d6b954bf7cdd922/e3VKfs7lB4tdDSzf0pNvz.png)
## Sampler Settings
The above examples were taken from the model using the below sampler settings. Feel free to experiment.
* Temp: 0.85
* TopK: 40
* TopP: 0.9
* RepPen: 1.1
## For Roleplay, use ChatML with System prompt, you may append the character name to the start messages if needed.
Formatting Settings for Silly Tavern. It should end up as ChatML with username/character names.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64dd7cda3d6b954bf7cdd922/AVkA54JFq-9avTMzqyDWx.png)
|