AWQ-Marlin Quantization Model

This model is based on the AWQ-Marlin quantization technique and is optimized for role-playing scenarios. It has been trained to generate dialogue and interactions in text-based RPGs. The model provides efficient performance without compromising on quality.

Use Cases

  • Role-playing Games (RPGs): Generate character dialogues and plot interactions in text-based RPGs.
Downloads last month
6
Safetensors
Model size
1.98B params
Tensor type
I32
·
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.