π«π· Calme-2
Collection
New Calme-2 fine-tuned models
β’
30 items
β’
Updated
β’
4
This is a fine-tuned version of the Qwen/Qwen2-7B
model. It aims to improve the base model across all benchmarks.
All GGUF models are available here: MaziyarPanahi/calme-2.1-qwen2-7b
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 23.20 |
IFEval (0-Shot) | 38.16 |
BBH (3-Shot) | 31.01 |
MATH Lvl 5 (4-Shot) | 21.07 |
GPQA (0-shot) | 5.26 |
MuSR (0-shot) | 13.80 |
MMLU-PRO (5-shot) | 29.92 |
This model uses ChatML
prompt template:
<|im_start|>system
{System}
<|im_end|>
<|im_start|>user
{User}
<|im_end|>
<|im_start|>assistant
{Assistant}
# Use a pipeline as a high-level helper
from transformers import pipeline
messages = [
{"role": "user", "content": "Who are you?"},
]
pipe = pipeline("text-generation", model="MaziyarPanahi/calme-2.1-qwen2-7b")
pipe(messages)
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-2.1-qwen2-7b")
model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-2.1-qwen2-7b")