QuantFactory/Violet_Twilight-v0.2-GGUF
This is quantized version of Epiculous/Violet_Twilight-v0.2 created using llama.cpp
Original Model Card
Now for something a bit different, Violet_Twilight-v0.2! This model is a SLERP merge of Azure_Dusk-v0.2 and Crimson_Dawn-v0.2!
Quants!
Prompting
The v0.2 models are trained on ChatML, the prompting structure goes a little something like this:
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
Context and Instruct
The v0.2 models are trained on ChatML, please use that Context and Instruct template.
Current Top Sampler Settings
Spicy_Temp
Violet_Twilight-Nitral-Special
Merging
The following config was used to merge Azure Dusk and Crimson Dawn
slices:
- sources:
- model: Epiculous/Azure_Dusk-v0.2
layer_range: [0, 40]
- model: Epiculous/Crimson_Dawn-V0.2
layer_range: [0, 40]
merge_method: slerp
base_model: Epiculous/Azure_Dusk-v0.2
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5 # fallback for rest of tensors
dtype: bfloat16
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 18.53 |
IFEval (0-Shot) | 45.32 |
BBH (3-Shot) | 23.94 |
MATH Lvl 5 (4-Shot) | 2.72 |
GPQA (0-shot) | 2.13 |
MuSR (0-shot) | 13.61 |
MMLU-PRO (5-shot) | 23.45 |
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 18.53 |
IFEval (0-Shot) | 45.32 |
BBH (3-Shot) | 23.94 |
MATH Lvl 5 (4-Shot) | 2.72 |
GPQA (0-shot) | 2.13 |
MuSR (0-shot) | 13.61 |
MMLU-PRO (5-shot) | 23.45 |
- Downloads last month
- 187
Datasets used to train QuantFactory/Violet_Twilight-v0.2-GGUF
Evaluation results
- strict accuracy on IFEval (0-Shot)Open LLM Leaderboard45.320
- normalized accuracy on BBH (3-Shot)Open LLM Leaderboard23.940
- exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard2.720
- acc_norm on GPQA (0-shot)Open LLM Leaderboard2.130
- acc_norm on MuSR (0-shot)Open LLM Leaderboard13.610
- accuracy on MMLU-PRO (5-shot)test set Open LLM Leaderboard23.450