cgato commited on
Commit
3a26373
·
verified ·
1 Parent(s): 1b53c31

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -4,6 +4,8 @@ license: cc-by-nc-4.0
4
  ## There's some kind of gguf tokenization issue where its not properly handling the <|im_end|> EOS token and instead breaks it into multiple smaller ones which degrades performance for gguf users. Need to submit an issue ticket. Unsure if this impacts other quantization formats.
5
  ## If possible, please use the safetensors or Exllama2 for best performance. Exllama2 tokenizes correctly.
6
 
 
 
7
  This is the first public release of the Humanize model I've been working on. That way folks have a stable repo to download from and can expect some consistient level of performance when downloading.
8
  ## Goals
9
  The goal of this model is to make something that feels different from other models out there. I wanted to accomplish this in two ways, first use as little synthetic data as possible, and second, by hand judging KTO samples to avoid common AI phrases and repetition. It has seen very very little synthetic data and without RL is unusable. The RL dataset was built by pulling responses from the model itself and then having a human (me) review each response.
 
4
  ## There's some kind of gguf tokenization issue where its not properly handling the <|im_end|> EOS token and instead breaks it into multiple smaller ones which degrades performance for gguf users. Need to submit an issue ticket. Unsure if this impacts other quantization formats.
5
  ## If possible, please use the safetensors or Exllama2 for best performance. Exllama2 tokenizes correctly.
6
 
7
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
8
+
9
  This is the first public release of the Humanize model I've been working on. That way folks have a stable repo to download from and can expect some consistient level of performance when downloading.
10
  ## Goals
11
  The goal of this model is to make something that feels different from other models out there. I wanted to accomplish this in two ways, first use as little synthetic data as possible, and second, by hand judging KTO samples to avoid common AI phrases and repetition. It has seen very very little synthetic data and without RL is unusable. The RL dataset was built by pulling responses from the model itself and then having a human (me) review each response.