--- base_model: - IntervitensInc/Mistral-Nemo-Base-2407-chatml - MarinaraSpaghetti/NemoMix-Unleashed-12B - inflatebot/MN-12B-Mag-Mell-R1 - LatitudeGames/Wayfarer-12B - PocketDoc/Dans-PersonalityEngine-V1.1.0-12b - TheDrummer/UnslopNemo-12B-v4 - yuyouyu/Mistral-Nemo-BD-RP - rAIfle/Questionable-MN-bf16 - romaingrx/red-teamer-mistral-nemo - crestf411/MN-Slush - aixonlab/Zinakha-12b - benhaotang/nemo-math-science-philosophy-12B - ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.2 library_name: transformers tags: - mergekit - merge - 12b - chat - roleplay - creative-writing - model-stock license: apache-2.0 --- # wuriaee-12B-schizostock > [*That.* is crazy.](https://youtu.be/zEWJ-JgVS7Q) This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). This is a merge of 14 models that I found interesting. I've downloaded them all and plan to make proper merges with them. Before I could make hypothetical good combinations, I thought it'd be funny to merge them all via model stock. I did give some thought to the order of the models in the config. More interesting models are at the top of the list in the full config and more stable ones are towards the bottom. The results were interesting, indeed. Haven't tested it much at all, but the results were intriguing. Tenth model. ## Merge Details ### Merge Method This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [IntervitensInc/Mistral-Nemo-Base-2407-chatml](https://huggingface.co./IntervitensInc/Mistral-Nemo-Base-2407-chatml) as a base. ### Models Merged The following models were included in the merge: * [MarinaraSpaghetti/NemoMix-Unleashed-12B](https://huggingface.co./MarinaraSpaghetti/NemoMix-Unleashed-12B) * [ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.2](https://huggingface.co./ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.2) * [inflatebot/MN-12B-Mag-Mell-R1](https://huggingface.co./inflatebot/MN-12B-Mag-Mell-R1) * [DavidAU/MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS](https://huggingface.co./DavidAU/MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS) * [yuyouyu/Mistral-Nemo-BD-RP](https://huggingface.co./yuyouyu/Mistral-Nemo-BD-RP) * [crestf411/MN-Slush](https://huggingface.co./crestf411/MN-Slush) * [IntervitensInc/Mistral-Nemo-Base-2407-chatml](https://huggingface.co./IntervitensInc/Mistral-Nemo-Base-2407-chatml) * [LatitudeGames/Wayfarer-12B](https://huggingface.co./LatitudeGames/Wayfarer-12B) * [PocketDoc/Dans-PersonalityEngine-V1.1.0-12b](https://huggingface.co./PocketDoc/Dans-PersonalityEngine-V1.1.0-12b) * [benhaotang/nemo-math-science-philosophy-12B](https://huggingface.co./benhaotang/nemo-math-science-philosophy-12B) * [aixonlab/Zinakha-12b](https://huggingface.co./aixonlab/Zinakha-12b) * [rAIfle/Questionable-MN-bf16](https://huggingface.co./rAIfle/Questionable-MN-bf16) * [romaingrx/red-teamer-mistral-nemo](https://huggingface.co./romaingrx/red-teamer-mistral-nemo) ### Configuration The following YAML configurations were used to produce this model: ```yaml # Full Configuration models: # Next 4 models are p1; the base model, tokenizer, template is the same for each part (same as the final merge) - model: inflatebot/MN-12B-Mag-Mell-R1 - model: LatitudeGames/Wayfarer-12B - model: PocketDoc/Dans-PersonalityEngine-V1.1.0-12b - model: TheDrummer/UnslopNemo-12B-v4 # p2: - model: yuyouyu/Mistral-Nemo-BD-RP - model: rAIfle/Questionable-MN-bf16 - model: romaingrx/red-teamer-mistral-nemo - model: MarinaraSpaghetti/NemoMix-Unleashed-12B # p3: - model: crestf411/MN-Slush - model: aixonlab/Zinakha-12b - model: benhaotang/nemo-math-science-philosophy-12B - model: ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.2 base_model: IntervitensInc/Mistral-Nemo-Base-2407-chatml merge_method: model_stock dtype: bfloat16 chat_template: "chatml" tokenizer: source: union ``` ```yaml # Final Model: models: - model: p1 - model: p2 - model: p3 base_model: IntervitensInc/Mistral-Nemo-Base-2407-chatml merge_method: model_stock dtype: bfloat16 chat_template: "chatml" tokenizer: source: union ```