--- license: cc-by-nc-4.0 tags: - not-for-all-audiences - nsfw model-index: - name: Mixtral-8x7B-MoE-RP-Story results: - task: type: text-generation name: Text Generation dataset: name: AI2 Reasoning Challenge (25-Shot) type: ai2_arc config: ARC-Challenge split: test args: num_few_shot: 25 metrics: - type: acc_norm value: 51.54 name: normalized accuracy source: url: https://huggingface.co./spaces/open-llm-leaderboard/open_llm_leaderboard?query=Undi95/Mixtral-8x7B-MoE-RP-Story name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: HellaSwag (10-Shot) type: hellaswag split: validation args: num_few_shot: 10 metrics: - type: acc_norm value: 70.0 name: normalized accuracy source: url: https://huggingface.co./spaces/open-llm-leaderboard/open_llm_leaderboard?query=Undi95/Mixtral-8x7B-MoE-RP-Story name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MMLU (5-Shot) type: cais/mmlu config: all split: test args: num_few_shot: 5 metrics: - type: acc value: 43.04 name: accuracy source: url: https://huggingface.co./spaces/open-llm-leaderboard/open_llm_leaderboard?query=Undi95/Mixtral-8x7B-MoE-RP-Story name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: TruthfulQA (0-shot) type: truthful_qa config: multiple_choice split: validation args: num_few_shot: 0 metrics: - type: mc2 value: 41.53 source: url: https://huggingface.co./spaces/open-llm-leaderboard/open_llm_leaderboard?query=Undi95/Mixtral-8x7B-MoE-RP-Story name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: Winogrande (5-shot) type: winogrande config: winogrande_xl split: validation args: num_few_shot: 5 metrics: - type: acc value: 67.32 name: accuracy source: url: https://huggingface.co./spaces/open-llm-leaderboard/open_llm_leaderboard?query=Undi95/Mixtral-8x7B-MoE-RP-Story name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: GSM8k (5-shot) type: gsm8k config: main split: test args: num_few_shot: 5 metrics: - type: acc value: 9.93 name: accuracy source: url: https://huggingface.co./spaces/open-llm-leaderboard/open_llm_leaderboard?query=Undi95/Mixtral-8x7B-MoE-RP-Story name: Open LLM Leaderboard --- Mixtral-8x7B-MoE-RP-Story is a model made primarely for chatting, RP (Roleplay) and storywriting. 2 RP model, 2 chat model, 1 occult model, 1 storywritting model, 1 mathematic model and 1 DPO model was used for a MoE. Bagel was the base. The DPO chat model is here to help get more human reply. This is my first try at doing this, so don't hesitate to give feedback! WARNING: ALL THE "K" GGUF QUANT OF MIXTRAL MODELS SEEMS TO BE [BROKEN](https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/TvjEP14ps7ZUgJ-0-mhIX.png), PREFER Q4_0, Q5_0 or Q8_0! ## Description This repo contains fp16 files of Mixtral-8x7B-MoE-RP-Story. ## Models used The list of model used and their activator/theme can be found [here](https://huggingface.co./Undi95/Mixtral-8x7B-MoE-RP-Story/blob/main/config.yaml) ## Prompt template: Custom Using Bagel as a base let us a lot of different prompting system theorically, you can see all the prompting available [here](https://huggingface.co./jondurbin/bagel-7b-v0.1#prompt-formatting). If you want to support me, you can [here](https://ko-fi.com/undiai). # [Open LLM Leaderboard Evaluation Results](https://huggingface.co./spaces/open-llm-leaderboard/open_llm_leaderboard) Detailed results can be found [here](https://huggingface.co./datasets/open-llm-leaderboard/details_Undi95__Mixtral-8x7B-MoE-RP-Story) | Metric |Value| |---------------------------------|----:| |Avg. |47.23| |AI2 Reasoning Challenge (25-Shot)|51.54| |HellaSwag (10-Shot) |70.00| |MMLU (5-Shot) |43.04| |TruthfulQA (0-shot) |41.53| |Winogrande (5-shot) |67.32| |GSM8k (5-shot) | 9.93|