the model size of mlx-community/Qwen1.5-0.5B-Chat-4bit

#1
by AlexWuKing - opened
MLX Community org
edited Jul 15

I use this mlx command to convert the model to 4bits:
python -m mlx_lm.convert --hf-path Qwen/Qwen1.5-0.5B-Chat -q --q-bits 4
the result shows that the model.safetensors is 500+M size, but mlx-community/Qwen1.5-0.5B-Chat-4bit only got 261M

Is there any difference ?

Sign up or log in to comment