--- base_model: - Sao10K/L3-8B-Niitama-v1 - OEvortex/Emotional-llama-8B - ArliAI/ArliAI-Llama-3-8B-Formax-v1.0 - nothingiisreal/L3-8B-Celeste-V1.2 - ResplendentAI/Nymph_8B - TheDrummer/Llama-3SOME-8B-v2 - nothingiisreal/L3-8B-Instruct-Abliterated-DWP library_name: transformers tags: - mergekit - merge - not-for-all-audiences --- God this sucked to make, I'm getting burnout with this fucker so I'm releasing what I have as a partial update. I'll be focusing on L3.1 for the time being and taking a break from this project for a bit. What's done is... fine, I'm not super happy about it but it's objectively better then v0.1. ### Quants [GGUFs](https://huggingface.co./mradermacher/L3-Horizon-Anteros-v0.1.5-13B-GGUF) by [mradermacher](https://huggingface.co./mradermacher) [GGUFs](https://huggingface.co./backyardai/L3-Horizon-Anteros-v0.1.5-13B-GGUF) by [BackyardAI](https://huggingface.co./backyardai) ### Details & Recommended Settings (Still testing; subject to change) Very experimental, expect bugs. Thrives at more story heavy and narrative RP yet still excels with the basics like usual. Way more horny this time, no idea why. Slightly worse instruct following compared to v0.1. Really needs examples to fuction, otherwise it'll spit out garble easily. I tried to also curb l3's tendency to double line (leaving a space between paragraphs) to mild success. Calmer writing style. Has a certain tendency to speak for the {user}, but that's easily negated with a few instructs. Rec. Settings: ``` Template: Plain Text or L3 Temperature: 1.3 Min P: 0.1 Repeat Penalty: 1.05 Repeat Penalty Tokens: 256 ``` ### Models Merged & Merge Theory The following models were included in the merge: * [ResplendentAI/Nymph_8B](https://huggingface.co./ResplendentAI/Nymph_8B) * [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co./TheDrummer/Llama-3SOME-8B-v2) * [nothingiisreal/L3-8B-Instruct-Abliterated-DWP](https://huggingface.co./nothingiisreal/L3-8B-Instruct-Abliterated-DWP) * [Sao10K/L3-8B-Niitama-v1](https://huggingface.co./Sao10K/L3-8B-Niitama-v1) * [OEvortex/Emotional-llama-8B](https://huggingface.co./OEvortex/Emotional-llama-8B) * [ArliAI/ArliAI-Llama-3-8B-Formax-v1.0](https://huggingface.co./ArliAI/ArliAI-Llama-3-8B-Formax-v1.0) * [nothingiisreal/L3-8B-Celeste-V1.2](https://huggingface.co./nothingiisreal/L3-8B-Celeste-V1.2) can't be asked rn ### Config ```yaml slices: - sources: - layer_range: [0, 6] #6 model: ArliAI/ArliAI-Llama-3-8B-Formax-v1.0 - sources: - layer_range: [4, 8] #10 model: nothingiisreal/L3-8B-Celeste-V1 parameters: scale: - filter: q_proj value: [1, 0.89] - sources: - layer_range: [8, 12] #14 model: nothingiisreal/L3-8B-Instruct-Abliterated-DWP parameters: scale: - filter: k_proj value: [0.89, 1] - sources: - layer_range: [6, 12] #20 model: ArliAI/ArliAI-Llama-3-8B-Formax-v1.0 - sources: - layer_range: [10, 14] #24 model: nothingiisreal/L3-8B-Celeste-V1 - sources: - layer_range: [12, 16] #28 model: ArliAI/ArliAI-Llama-3-8B-Formax-v1.0 - sources: - layer_range: [14, 28] #42 model: nothingiisreal/L3-8B-Instruct-Abliterated-DWP - sources: - layer_range: [21, 31] #52 model: nothingiisreal/L3-8B-Celeste-V1 - sources: - layer_range: [28, 32] #56 model: nothingiisreal/L3-8B-Instruct-Abliterated-DWP parameters: int8_mask: true merge_method: passthrough dtype: float32 out_dtype: bfloat16 name: anterosv2.b --- slices: - sources: - layer_range: [0, 4] #4 model: ResplendentAI/Nymph_8B - sources: - layer_range: [2, 10] #12 model: Sao10K/L3-8B-Niitama-v1 - sources: - layer_range: [8, 12] #16 model: TheDrummer/Llama-3SOME-8B-v2 - sources: - layer_range: [10, 16] #22 model: Sao10K/L3-8B-Niitama-v1 - sources: - layer_range: [12, 20] #30 model: ResplendentAI/Nymph_8B - sources: - layer_range: [14, 18] #34 model: OEvortex/Emotional-llama-8B - sources: - layer_range: [16, 20] #38 model: ResplendentAI/Nymph_8B - sources: - layer_range: [18, 22] #42 model: OEvortex/Emotional-llama-8B - sources: - layer_range: [20, 24] #46 model: TheDrummer/Llama-3SOME-8B-v2 - sources: - layer_range: [23, 31] #54 model: ResplendentAI/Nymph_8B - sources: - layer_range: [30, 32] #56 model: Sao10K/L3-8B-Niitama-v1 parameters: int8_mask: true merge_method: passthrough dtype: float32 out_dtype: bfloat16 name: anterosv2.c --- slices: - sources: - layer_range: [2, 17] model: anterosv2.b parameters: int8_mask: true merge_method: passthrough dtype: float32 out_dtype: bfloat16 name: 2-17av2b.sl --- slices: - sources: - layer_range: [2, 17] model: anterosv2.c parameters: int8_mask: true merge_method: passthrough dtype: float32 out_dtype: bfloat16 name: 2-17av2c.sl --- models: - model: 2-17av2c.sl parameters: weight: [0.1, 0.5, 0.3] - model: 2-17av2b.sl parameters: weight: [0.9, 0.5, 0.7] merge_method: dare_linear base_model: 2-17av2b.sl parameters: normalize: false int8_mask: true dtype: float32 out_dtype: bfloat16 name: 2-17aav2.sl --- slices: - sources: - layer_range: [17, 42] model: anterosv2.b parameters: int8_mask: true merge_method: passthrough dtype: float32 out_dtype: bfloat16 name: 17-42av2b.sl --- slices: - sources: - layer_range: [17, 42] model: anterosv2.c parameters: int8_mask: true merge_method: passthrough dtype: float32 out_dtype: bfloat16 name: 17-42av2c.sl --- models: - model: 17-42av2b.sl parameters: weight: [0.8, 0.5, 0.4, 0.3, 0.15] density: 0.65 epsilon: 0.07 lambda: 0.12 - model: 17-42av2c.sl parameters: weight: [0.2, 0.5, 0.6, 0.7, 0.85] density: 0.7 epsilon: 0.05 lambda: 0.1 merge_method: della base_model: 17-42av2c.sl parameters: normalize: false int8_mask: true dtype: float32 out_dtype: bfloat16 name: 17-42aav2.sl --- slices: - sources: - layer_range: [42, 52] model: anterosv2.b parameters: int8_mask: true merge_method: passthrough dtype: float32 out_dtype: bfloat16 name: 42-52av2b.sl --- slices: - sources: - layer_range: [42, 52] model: anterosv2.c parameters: int8_mask: true merge_method: passthrough dtype: float32 out_dtype: bfloat16 name: 42-52av2c.sl --- models: - model: 42-52av2c.sl parameters: weight: [0.9, 0.65, 0.9] - model: 42-52av2b.sl parameters: weight: [0.1, 0.35, 0.1] merge_method: dare_linear base_model: 42-52av2c.sl parameters: normalize: false int8_mask: true dtype: float32 out_dtype: bfloat16 name: 42-52aav2.sl --- slices: - sources: - layer_range: [0, 2] model: anterosv2.b - sources: - layer_range: [0, 15] model: 2-17aav2.sl - sources: - layer_range: [0, 25] model: 17-42aav2.sl - sources: - layer_range: [0, 10] model: 42-52aav2.sl - sources: - layer_range: [52, 56] model: anterosv2.c parameters: int8_mask: true merge_method: passthrough dtype: float32 out_dtype: bfloat16 name: anterosv0.1.5 ```