Posting these Qwen-14B-Chat quantized models in GGUF format for use with llama.cpp due to a user request.

But, having used an importance matrix derived from English-only training data in the quantization, I have no idea how these models will perform in Chinese.

Downloads last month
31
GGUF
Model size
14.2B params
Architecture
qwen
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.