gemma-2-9B-it-iq1_m
This is a quantized version of the Gemma2 9B instruct model using the IQ1_M quantization method.
Model Details
- Original Model: Gemma2-9B-it
- Quantization Method: IQ1_M
- Precision: 1-bit
- iMatrix: From bartowski. You can find the file in the following repo; gemma-2-9b-it-gguf repo
Usage
You can use it directly with llama.cpp
- Downloads last month
- 0