YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co./docs/hub/model-cards#model-card-metadata)
DeepSeek-R1-Distill-Qwen-7B-q4f16_ft-MLC
Model Configuration | |
---|---|
Source Model | deepseek-ai/DeepSeek-R1-Distill-Qwen-7B |
Inference API | MLC_LLM |
Quantization | q4f16_ft |
Model Type | qwen2 |
Vocab Size | 152064 |
Context Window Size | 131072 |
Prefill Chunk Size | 8192 |
Temperature | 0.6 |
Repetition Penalty | 1.0 |
top_p | 0.95 |
pad_token_id | 0 |
bos_token_id | 151646 |
eos_token_id | 151643 |
See jetson-ai-lab.com/models.html
for benchmarks, examples, and containers to deploy local serving and inference for these quantized models.
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
HF Inference API was unable to determine this model's library.