DeepSeek-R1-Distill-Qwen-32B-GGUF
Original Model
deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
Run with LlamaEdge
LlamaEdge version: v0.16.1
Prompt template
Prompt type:
chatml
Prompt string
<|im_start|>system {system_message}<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant
Context size:
128000
Run as LlamaEdge service
wasmedge --dir .:. --nn-preload default:GGML:AUTO:DeepSeek-R1-Distill-Qwen-32B-Q5_K_M.gguf \ llama-api-server.wasm \ --model-name DeepSeek-R1-Distill-Qwen-32B \ --prompt-template chatml \ --ctx-size 128000
Run as LlamaEdge command app
wasmedge --dir .:. --nn-preload default:GGML:AUTO:DeepSeek-R1-Distill-Qwen-32B-Q5_K_M.gguf \ llama-chat.wasm \ --prompt-template chatml \ --ctx-size 128000
Quantized with llama.cpp b4519
- Downloads last month
- 211
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for second-state/DeepSeek-R1-Distill-Qwen-32B-GGUF
Base model
deepseek-ai/DeepSeek-R1-Distill-Qwen-32B