File size: 2,733 Bytes
ad62d0e 517387a ad62d0e 47c1bdb ad62d0e 47c1bdb c8f5376 953d798 47c1bdb 953d798 47c1bdb 953d798 47c1bdb 57ffa92 47c1bdb 57ffa92 47c1bdb ad62d0e 47c1bdb ad62d0e 47c1bdb ad62d0e 57ffa92 ad62d0e 21ef10c 47c1bdb 21ef10c 085f0f3 21ef10c 085f0f3 21ef10c 085f0f3 21ef10c ad62d0e 21ef10c ad62d0e 47c1bdb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 |
---
base_model:
- meta-llama/Llama-3.1-8B
- Hastagaras/snovalite-baukit-6-14.FT-L5-7.13-22.27-31
library_name: transformers
license: llama3.1
pipeline_tag: text-generation
tags:
- mergekit
- merge
- not-for-all-audiences
---
### ZABUZA
GGUF: [IMATRIX](https://huggingface.co./mradermacher/Zabuza-8B-Llama-3.1-i1-GGUF)/[STATIC](https://huggingface.co./mradermacher/Zabuza-8B-Llama-3.1-GGUF) made available by [mradermacher](https://huggingface.co./mradermacher)
---
This model is a combination of merge, abliteration technique (using baukit) and finetuning.
The base model is [arcee-ai/Llama-3.1-SuperNova-Lite](https://huggingface.co./arcee-ai/Llama-3.1-SuperNova-Lite), which underwent abliteration to reduce model refusals.
Next, I finetuned the abliterated SuperNova-Lite with 10K diverse examples such as:
* **Claude and Gemini Instruction/RP** (15k sloppy examples were removed!, but some may have slipped through.)
* **Human-written Stories/RP** (Most stories have dialogue)
* **IFEval-like data** (To preserve the model's instruction following ability)
* **Harmful data** (To remove disclaimers and moralizing responses, but not 100% disappear.)
* **My sarcastic and rude AI assistant data** (Just for my personal satisfaction)
Lastly, I merged the model using TIES, inspired by this [MERGE](https://huggingface.co./Joseph717171/Llama-3.1-SuperNova-8B-Lite_TIES_with_Base) by Joseph717171.
### Chat Template
Llama 3.1 Instruct
```
<|start_header_id|>{role}<|end_header_id|>
{message}<|eot_id|><|start_header_id|>{role}<|end_header_id|>
{message}<|eot_id|>
```
System messages for role-playing should be very detailed if you don't want dry responses.
### Example Output
This example was taken with the system prompt: `You're a rude and sarcastic AI assistant.`
<div align="left">
<img src="https://huggingface.co./Hastagaras/Zabuza-8B-Llama-3.1/resolve/main/1.png" width="1000"/>
</div>
<div align="left">
<img src="https://huggingface.co./Hastagaras/Zabuza-8B-Llama-3.1/resolve/main/2.png" width="1000"/>
</div>
<div align="left">
<img src="https://huggingface.co./Hastagaras/Zabuza-8B-Llama-3.1/resolve/main/3.png" width="1000"/>
</div>
### Configuration
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
The following YAML configuration was used to produce this model:
```yaml
models:
- model: Hastagaras/snovalite-baukit-6-14.FT-L5-7.13-22.27-31
parameters:
weight: 1
density: 1
- model: Hastagaras/snovalite-baukit-6-14.FT-L5-7.13-22.27-31
parameters:
weight: 1
density: 1
merge_method: ties
base_model: meta-llama/Llama-3.1-8B
parameters:
density: 1
normalize: true
int8_mask: true
dtype: bfloat16
``` |