File size: 1,589 Bytes
7846abe
ea31edc
 
 
 
7846abe
ea31edc
 
 
 
 
 
 
 
 
 
 
 
 
 
7846abe
ea31edc
 
 
 
 
 
 
 
 
81e077d
 
ea31edc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
---
inference: false
language:
- zh
- en
license: unknown
model_name: Rain-2x7B-MoE-32k-v0.2
pipeline_tag: text-generation
prompt_template: '<s> SYS_PROMPT [INST] QUERY1 [/INST] RESPONSE1 [INST] QUERY2 [/INST]'
tags:
- nlp
- chinese
- mistral
- mixtral
- traditional_chinese
- merge
- mergekit
- MediaTek-Research/Breeze-7B-Instruct-v0_1
- beowolx/CodeNinja-1.0-OpenChat-7B
- mlabonne/Marcoro14-7B-slerp
---

<br/>

# 小雨同學 2x7B

採用聯發科 Breeze 7B Instruct 為基底的國語 MoE (Mixture-of-Experts) 模型,共有兩個 Expert model。

請用 Marcoro14-7B 或是 Breeze-7B-Instruct 所推薦的 Prompt 格式進行操作;以下為模型配置。

- v0.2 更新了 tokenizer parameters

![](https://i.imgur.com/f3Ro6Fu.png)

### Rain-2x7B-MoE-32k-v0.2

This is an experimental Mixtral-architecture MoE model with 2 of 7B sized fine-tunes. Breeze and CodeNinja are used on top of [Marcoro14-7B-slerp](https://huggingface.co./mlabonne/Marcoro14-7B-slerp).

Model configuration is as follows:

* [Marcoro14-7B-slerp](https://huggingface.co./mlabonne/Marcoro14-7B-slerp) as base.
* [Breeze-7B-Instruct-v0_1](https://huggingface.co./MediaTek-Research/Breeze-7B-Instruct-v0_1) as model 0.
* [CodeNinja-1.0-OpenChat-7B](https://huggingface.co./beowolx/CodeNinja-1.0-OpenChat-7B) as model 1.

To use the model, please use either prompt templates suggested by the base models.

## Notes

Please evaluate before use in any application pipeline. Activation for coding part of the model would be `'code'`, `'python'`, `'typescript'`, `'javascript'`, `'programming'`, `'algorithm'`.