BitNet Q2B_X
Collection
Models Quantized with Q2B0
•
8 items
•
Updated
The Llama3-8B-1.58-100B-tokens-q2b0 is a quantized version of Llama3-8B-1.58-100B-tokens-GGUF, leveraging the q2b0 quantization method from Candle. This enables extreme compression while maintaining strong performance across various NLP tasks.
The model has been quantized using the q2b0 method from Candle. This approach reduces model size significantly while preserving performance. For more details on this quantization technique, refer to the Candle PR #2683.
For details on the dataset and training process, refer to the original Llama3-8B-1.58-100B-tokens.
Base model
meta-llama/Meta-Llama-3-8B-Instruct