If there will be a gguf version?
#1
by
Cran-May
- opened
Now LLaMA.cpp supports Qwen
https://github.com/ggerganov/llama.cpp/pull/4281
Here is a series of qwen models.(The support of 1.8B model is not checked.)
https://huggingface.co./Qwen/Qwen-72B-Chat
https://huggingface.co./Qwen/Qwen-14B-Chat
https://huggingface.co./Qwen/Qwen-7B-Chat
https://huggingface.co./Qwen/Qwen-1_8B-Chat
https://huggingface.co./Qwen/Qwen-Audio-Chat
https://huggingface.co./Qwen/Qwen-VL-Chat
Maybe Internlm? https://github.com/ggerganov/llama.cpp/pull/4283
https://huggingface.co./internlm/internlm-chat-20b
https://huggingface.co./internlm/internlm-chat-7b-v1_1
Qwen-VL Chat seems not to be supported.
This comment has been hidden
This comment has been hidden