GGUFs PLUS:

Q8 and Q6 GGUFs with critical parts of the model in F16 / Full precision.

File sizes will be slightly larger than standard, but should yeild higher quality results under all tasks and conditions.

Downloads last month
9
GGUF
Model size
10.7B params
Architecture
llama

6-bit

8-bit

Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Collection including DavidAU/LemonadeRP-4.5.3-11B-GGUF-Plus