🇫🇷 Calme-3
Collection
Here you can find all the new Calme-3 models
•
25 items
•
Updated
•
6
This is avery small model, so it might not perform well for some prompts and may be sensitive to hyper parameters. I would appreciate any feedback to see if I can fix any issues in the next iteration. ❤️
This model is an advanced iteration of the powerful Qwen/Qwen2.5-3B, fine-tuned specifically to enhance its capabilities across general domains in both French and English.
All GGUF models are available here: MaziyarPanahi/calme-3.2-baguette-3b-GGUF
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 22.14 |
IFEval (0-Shot) | 63.38 |
BBH (3-Shot) | 25.87 |
MATH Lvl 5 (4-Shot) | 3.10 |
GPQA (0-shot) | 5.93 |
MuSR (0-shot) | 8.60 |
MMLU-PRO (5-shot) | 25.98 |
This model uses ChatML
prompt template:
<|im_start|>system
{System}
<|im_end|>
<|im_start|>user
{User}
<|im_end|>
<|im_start|>assistant
{Assistant}
# Use a pipeline as a high-level helper
from transformers import pipeline
messages = [
{"role": "user", "content": "Who are you?"},
]
pipe = pipeline("text-generation", model="MaziyarPanahi/calme-3.2-baguette-3b")
pipe(messages)
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-3.2-baguette-3b")
model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-3.2-baguette-3b")
As with any large language model, users should be aware of potential biases and limitations. We recommend implementing appropriate safeguards and human oversight when deploying this model in production environments.