bethrezen's picture
add - [Q4_K_M GGUF](https://huggingface.co./ZeroAgency/Zero-Mistral-Small-24B-Instruct-2501-Q4_K_M)
93a34d0
---
license: mit
datasets:
- Vikhrmodels/GrandMaster-PRO-MAX
language:
- en
- ru
tags:
- mistral
- chat
- conversational
- transformers
inference:
parameters:
temperature: 0
pipeline_tag: text-generation
base_model:
- mistralai/Mistral-Small-24B-Instruct-2501
- ZeroAgency/Zero-Mistral-Small-24B-Instruct-2501
library_name: vllm
---
# Zero-Mistral-Small-24B-Instruct-2501 Q8_0 GGUF version
Zero-Mistral-Small is an improved version of [mistralai/Mistral-Small-24B-Instruct-2501](https://huggingface.co./mistralai/Mistral-Small-24B-Instruct-2501), primarily adapted for Russian and English languages.
The training involved SFT stage on [GrandMaster-PRO-MAX](https://huggingface.co./datasets/Vikhrmodels/GrandMaster-PRO-MAX) dataset.
## ๐Ÿ“š Model versions
- [Merged 16-bit](https://huggingface.co./ZeroAgency/Zero-Mistral-Small-24B-Instruct-2501) - original 16bit merged version.
- [LoRa adapter](https://huggingface.co./ZeroAgency/Zero-Mistral-Small-24B-Instruct-2501-lora) for mistralai/Mistral-Small-24B-Instruct-2501
- [F16 GGUF](https://huggingface.co./ZeroAgency/Zero-Mistral-Small-24B-Instruct-2501-F16)
- [BF16 GGUF](https://huggingface.co./ZeroAgency/Zero-Mistral-Small-24B-Instruct-2501-BF16)
- [Q8_0 GGUF](https://huggingface.co./ZeroAgency/Zero-Mistral-Small-24B-Instruct-2501-Q8_0)
- [Q4_K_M GGUF](https://huggingface.co./ZeroAgency/Zero-Mistral-Small-24B-Instruct-2501-Q4_K_M)