Kquant03 commited on
Commit
9b7ae80
1 Parent(s): e898346

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -0
README.md CHANGED
@@ -11,6 +11,13 @@ tags:
11
 
12
  A frankenMoE not only using far better methodology and fundamental understanding of SMoE, but completely focused around intellectual roleplay. It may have a bit of a redundancy issue, to battle this, try to keep things fresh with the model by either introducing new concepts often, or through [drμgs](https://github.com/EGjoni/DRUGS). (no not that kind)
13
 
 
 
 
 
 
 
 
14
  # "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
15
  ### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
16
 
 
11
 
12
  A frankenMoE not only using far better methodology and fundamental understanding of SMoE, but completely focused around intellectual roleplay. It may have a bit of a redundancy issue, to battle this, try to keep things fresh with the model by either introducing new concepts often, or through [drμgs](https://github.com/EGjoni/DRUGS). (no not that kind)
13
 
14
+ The config looks like this...(detailed version is in the files and versions):
15
+ - [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) - base
16
+ - [fblgit/una-cybertron-7b-v3-OMA](https://huggingface.co/fblgit/una-cybertron-7b-v3-OMA) - expert #1
17
+ - [rwitz/go-bruins-v2](https://huggingface.co/rwitz/go-bruins-v2) - expert #2
18
+ - [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) - expert #3
19
+ - [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) - expert #4
20
+
21
  # "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
22
  ### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
23