File size: 1,254 Bytes
1cae86c b36cd2b a3cbf7f 220e2bb 00d15ed 220e2bb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 |
---
license: other
license_name: sacla
license_link: >-
https://huggingface.co./stabilityai/stable-diffusion-3.5-large-turbo/blob/main/LICENSE.md
base_model:
- stabilityai/stable-diffusion-3.5-large-turbo
base_model_relation: quantized
---
These models are made to work with [stable-diffusion.cpp](https://github.com/leejet/stable-diffusion.cpp) release [master-ac54e00](https://github.com/leejet/stable-diffusion.cpp/releases/tag/master-ac54e00) onwards. Support for other inference backends is not guarenteed.
Quantized using this PR https://github.com/leejet/stable-diffusion.cpp/pull/447
### Files:
- [sd3.5_large_turbo-q2_k_4_0.gguf](https://huggingface.co./stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q2_k_4_0.gguf): Smallest quantization yet. Use this if you can't afford anything bigger
- [sd3.5_large_turbo-q4_k_4_0.gguf](https://huggingface.co./stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q4_k_4_0.gguf): Exacty same size as q4_0, but with slightly less degradation. Recommended
- [sd3.5_large_turbo-q4_k_4_1.gguf](https://huggingface.co./stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q4_k_4_1.gguf): Smaller than q4_1, and with comparable degradation. Recommended |