bjoernp commited on
Commit
629a1d3
1 Parent(s): 66c02d4

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +136 -0
README.md ADDED
@@ -0,0 +1,136 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - migtissera/Synthia-v1.3
4
+ - meta-math/MetaMathQA
5
+ - NousResearch/capybara
6
+ language:
7
+ - en
8
+ library_name: transformers
9
+ pipeline_tag: text-generation
10
+ model_creator: DiscoResearch
11
+ model_type: mixtral
12
+ Tags:
13
+ - mixtral
14
+ - moe
15
+ - discoresearch
16
+ ---
17
+
18
+
19
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/62e3b6ab0c2a907c388e4965/IP6ULgm4XLcK_JLRz-WV4.png)
20
+ *Eight french experts sitting at a table. There's lots of wind.*
21
+
22
+ # DiscoLM Mixtral 8x7b alpha
23
+
24
+ **DiscoLM Mixtral 8x7b alpha** is an experimental 8x7b MoE model based on [Mistral AI´s Mixtral 8x7b](https://twitter.com/MistralAI/status/1733150512395038967).
25
+ This model is based on experimental code converting the model weights to huggingface format and enabling Transformers-based inference.
26
+ It was then finetuned on the OpenHermes, MethaMathQA und Tested-22k-Python-Alpaca datasets.
27
+ DiscoLM Mixtral 8x7b alpha is a [DiscoResearch](https://huggingface.co/DiscoResearch) project and was created by [Björn Plüster](https://huggingface.co/bjoernp) with lots of support from the community.
28
+
29
+ **Many thanks to [HessianAI](https://hessian.ai/) for providing the compute resources for this project and to the great people at [LAION](https://laion.ai) without whom this project would not have been possible!**
30
+
31
+
32
+ ## Table of Contents
33
+
34
+ 1. [Download](#download)
35
+ 2. [Benchmarks](#benchmarks)
36
+ 3. [Prompt Format](#prompt-format)
37
+ 4. [Dataset](#dataset)
38
+ 5. [Acknowledgements](#acknowledgements)
39
+ 6. [Contact](#contact)
40
+ 7. [About DiscoResearch](#about-discoresearch)
41
+ 8. [Disclaimer](#disclaimer)
42
+
43
+ ## Download
44
+
45
+ **Please note that you have to run the model with `trust_remote_code=True` until the new arch is merged into transformers!**
46
+
47
+ | Huggingface | GPTQ | GGUF | AWQ | *Base Model* |
48
+ |-------|-------|-------|-------|-------|
49
+ | [Link](https://huggingface.co/DiscoResearch/DiscoLM-Mixtral-8x7b) | tbc | tbc | tbc | tbc |
50
+
51
+ ## Benchmarks
52
+
53
+ ### Huggingface Leaderboard
54
+
55
+ This model is still an early Alpha with experimental code and we can't guarantee that there all values are correct.
56
+ The following are the scores from our own evaluation.
57
+
58
+ | Metric | Value |
59
+ |-----------------------|-------|
60
+ | ARC (25-shot) | xx |
61
+ | HellaSwag (10-shot) | xx |
62
+ | MMLU (5-shot) | xx |
63
+ | TruthfulQA (0-shot) | xx |
64
+ | Winogrande (5-shot) | xx |
65
+ | GSM8k (5-shot) | xx |
66
+ | **Avg.** | **xx** |
67
+
68
+
69
+ We use [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) to run the benchmark tests above, using the same version as the HuggingFace LLM Leaderboard.
70
+
71
+ ### FastEval
72
+ tbc
73
+
74
+ ### MTBench
75
+
76
+ tbc
77
+
78
+ ## Prompt Format
79
+
80
+ **Please note that you have to run the model with `trust_remote_code=True` until the new arch is merged into transformers!**
81
+
82
+ This model follows the ChatML format:
83
+
84
+ ```
85
+ <|im_start|>system
86
+ You are DiscoLM, a helpful assistant.
87
+ <|im_end|>
88
+ <|im_start|>user
89
+ Please tell me possible reasons to call a research collective "Disco Research"<|im_end|>
90
+ <|im_start|>assistant
91
+ ```
92
+
93
+ This formatting is also available via a pre-defined Transformers chat template, which means that lists of messages can be formatted for you with the apply_chat_template() method:
94
+
95
+ ```python
96
+ chat = [
97
+ {"role": "system", "content": "You are DiscoLM, a helpful assistant."},
98
+ {"role": "user", "content": "Please tell me possible reasons to call a research collective Disco Research"}
99
+ ]
100
+ tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=True)
101
+ ```
102
+
103
+ If you use `tokenize=True` and `return_tensors="pt"` instead, then you will get a tokenized and formatted conversation ready to pass to `model.generate()`.
104
+
105
+ ## Datasets
106
+
107
+ The following datasets were used for training DiscoLM Mixtral 8x7b alpha:
108
+
109
+ * [Synthia](https://huggingface.co/datasets/migtissera/Synthia-v1.3)
110
+ * [MetaMathQA](https://huggingface.co/datasets/meta-math/MetaMathQA)
111
+ * NousReseach Capybara (currently not public)
112
+
113
+ Many thanks for all dataset providers/curators!
114
+
115
+ ## Contact
116
+
117
+ Best way to reach us is on our [Discord](https://discord.gg/S8W8B5nz3v).
118
+
119
+ ## About DiscoResearch
120
+
121
+ DiscoResearch is an aspiring open research community. Disco should be a place where researchers from many communities can come together to combine their expertise and create innovative and groundbreaking LLMs. Come join our Discord, share your opinions and ideas, and advance open LLM research with us!
122
+
123
+ ## Acknowledgements
124
+
125
+ Many thanks first and foremost to [Mistral AI](https://huggingface.co/mistralai) for releasing another awesome model and their release strategy that is much fun for the whole community.
126
+ Additionally, many thanks in particular to [Dmytro Dzhulgakov](https://huggingface.co/dzhulgakov) who was the first one with a running [inference implementation](https://github.com/dzhulgakov/llama-mistral), [Vik](https://huggingface.co/vikhyatk) who spotted a critical bug in our first implementation (he actually read the paper!), [winglian](https://huggingface.co/winglian) for helpful advice and Axolotl which was used to finetune the model, [Teknium](https://huggingface.co/teknium), [Verzora](https://huggingface.co/Vezora) [MetaMath](https://huggingface.co/meta-math) for their great datasets, and everyone who participated in this awesome speedrun on either our, the [Nous Research](https://huggingface.co/NousResearch) or one of the other Discords (please contact us if we forgot to mention you here!).
127
+
128
+ **DiscoLM Mixtral is a [DiscoResearch](https://huggingface.co/DiscoResearch) project and was created by [Björn Plüster](https://huggingface.co/bjoernp).
129
+ The model was trained with compute provided by [HessianAI](https://hessian.ai/); many thanks as well to [LAION](https://laion.ai) for their coordination and providing invaluable contacts + advice.**
130
+
131
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
132
+
133
+ ## Disclaimer
134
+
135
+ The license on this model does not constitute legal advice. We are not responsible for the actions of third parties who use this model.
136
+ This model should only be used for research purposes.