File size: 1,718 Bytes
dad3663 b25cc0b dad3663 b25cc0b 790abd5 fc48c50 b25cc0b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 |
---
license: llama2
tags:
- merge
- mergekit
---
I merged [Aurelian](https://huggingface.co./grimulkan/aurelian-v0.5-70b-rope8-32K-fp16) with itself using [mergekit](https://github.com/cg123/mergekit), creating this EXTENDED LENGTH FRANKENSTEIN.
# Does it work
Yes, at 17k it stays coherent, but starts to lose minor details of the story. Not sure how well it performs at 32k though. Quants have a **sinificant** impact on quality for this model, going from Q6_K to Q5_K had a noticeable drop in quality.
# Is it worth it
Maybe? Depends? Do you hate mixtral? Do you have good hardware/patience? Do you need a somewhat smart model with 32k context?
# Known issues
VERY strict adherence to prompt format, forgetfullness, strong roleplay bias.
# Personal opinion
Dumber than Goliath, but has much less GPTism. If you want 32k goliath, maybe try [Goliath-longLORA-120b-rope8-32k-fp16](https://huggingface.co./grimulkan/Goliath-longLORA-120b-rope8-32k-fp16).
# Prompt format
Same as [Aurelian 0.5](https://huggingface.co./grimulkan/aurelian-v0.5-70b-rope8-32K-fp16).
```
[INST] <<SYS>>
System prompt, default is: An interaction between a user providing instructions, and an imaginative assistant providing responses.
<</SYS>>
</s><s>[INST] {Put your input text here.}
[/INST] {Model output}
```
This model doesn't like it too much when you change the prompt format, so even keeping that ```</s><s>``` is important.
# Benchmarks
### NeoEvalPlusN_benchmark
[My meme benchmark.](https://huggingface.co./datasets/ChuckMcSneed/NeoEvalPlusN_benchmark)
| Test name | Aurelian | DoubleGold |
| ---------- | ---------- | ------- |
| B | 1 | 1 |
| C | 1 | 1 |
| D | 0 | 2 |
| S | 2.5 | 3.25 |
| P | 2.25 | 1.5 |
| Total | 6.75 | 8.75 | |