Upscaled Models ⏫
Collection
A collection of my frankenmerges, upscaling several models. All of them have the corresponding GGUF variants.
•
4 items
•
Updated
•
2
This is SOLAR-10.7B, but upscaled to 13B, to optimize VRAM usage of typical GPU cards (a 4bit quant fits in 12GB).
Evaluations coming soon!
This is a frankenmerge model created using mergekit.
This model was merged using the passthrough merge method.
The following models were included in the merge:
The following YAML configuration was used to produce this model:
slices:
- sources:
- model: upstage/SOLAR-10.7B-Instruct-v1.0
layer_range: [0, 28]
- sources:
- model: upstage/SOLAR-10.7B-Instruct-v1.0
layer_range: [20, 48]
merge_method: passthrough
dtype: float16
The same as in SOLAR-10.7B:
<s> ### User:
{prompt}
### Assistant:
{response}</s>
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 56.65 |
AI2 Reasoning Challenge (25-Shot) | 57.25 |
HellaSwag (10-Shot) | 78.03 |
MMLU (5-Shot) | 55.75 |
TruthfulQA (0-shot) | 61.99 |
Winogrande (5-shot) | 70.24 |
GSM8k (5-shot) | 16.60 |
Base model
upstage/SOLAR-10.7B-v1.0