--- base_model: v000000/L3.1-Celestial-Stone-2x8B-DPO library_name: transformers tags: - merge - llama - mixtral - dpo - llama-cpp --- # L3.1-Celestial-Stone-2x8B-DPO (GGUFs) This model was converted to GGUF format from [`v000000/L3.1-Celestial-Stone-2x8B-DPO`](https://huggingface.co./v000000/L3.1-Celestial-Stone-2x8B-DPO) using llama.cpp. Refer to the [original model card](https://huggingface.co./v000000/L3.1-Celestial-Stone-2x8B-DPO) for more details on the model. ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f74b6e6389380c77562762/PIC3kb7XL2f14YhLkrRsm.png) # Ordered by quality: * q8_0 imatrix --- 14.2g * q6_k imatrix --- 11.2g * q5_k_s imatrix --- 9.48g * iq4_xs imatrix --- 7.44g Missing? See [mradermacher i1](https://huggingface.co./mradermacher/L3.1-Celestial-Stone-2x8B-DPO-i1-GGUF) for more types of imatrix quants. imatrix data (V2 - 287kb) randomized bartowski, kalomeze groups, ERP/RP snippets, working gpt4 code, toxic qa, human messaging, randomized posts, story, novels