PrimaMonarch-EroSumika-2x10.7B
Collection
OUTDATED
•
4 items
•
Updated
•
1
(Maybe i'll change the icon picture later.)
Experimental MoE, the idea is to have more active parameters than 7xX model would have and keep it's size lower than 20B.
This model has ~19.2B parameters.
Exl2, 4.0 bpw (Fits in 12GB VRAM/16k context/4-bit cache)
slices:
- sources:
- model: MistralInstruct-v0.2-128k
layer_range: [0, 24]
- sources:
- model: MistralInstruct-v0.2-128k
layer_range: [8, 24]
- sources:
- model: MistralInstruct-v0.2-128k
layer_range: [24, 32]
merge_method: passthrough
dtype: bfloat16
xxx777xxxASD/PrimaSumika-10.7B-128k
slices:
- sources:
- model: EroSumika-128k
layer_range: [0, 24]
- sources:
- model: Prima-Lelantacles-128k
layer_range: [8, 24]
- sources:
- model: EroSumika-128k
layer_range: [24, 32]
merge_method: passthrough
dtype: bfloat16
slices:
- sources:
- model: AlphaMonarch-7B-128k
layer_range: [0, 24]
- sources:
- model: NeuralHuman-128k
layer_range: [8, 24]
- sources:
- model: AlphaMonarch-7B-128k
layer_range: [24, 32]
merge_method: passthrough
dtype: bfloat16
Each 128k model is a slerp merge with Epiculous/Fett-uccine-Long-Noodle-7B-120k-Context