Qwen2.5-0.5B-Instruct

Introduction

This model is based on the Qwen2.5-0.5B-Instruct model and is quantized in 4bits in the EXL2 format using the AutoQuant system : https://colab.research.google.com/drive/1b6nqC7UZVt8bx4MksX7s656GXPM-eWw4

You can learn more about the EXL2 format here : https://github.com/turboderp/exllamav2 Feel free to use it as you want

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for Volko76/Qwen2.5-0.5B-Instruct-EXL2-4bits

Base model

Qwen/Qwen2.5-0.5B
Finetuned
(95)
this model

Collection including Volko76/Qwen2.5-0.5B-Instruct-EXL2-4bits