--- license: apache-2.0 language: - en metrics: - accuracy base_model: - Qwen/Qwen2.5-Math-7B-Instruct library_name: transformers pipeline_tag: question-answering datasets: - MATH - GSM8K --- Quantization made by Richard Erkhov. [Github](https://github.com/RichardErkhov) [Discord](https://discord.gg/pvy7H8DZMG) [Request more models](https://github.com/RichardErkhov/quant_request) SuperCorrect-7B - GGUF - Model creator: https://huggingface.co./BitStarWalkin/ - Original model: https://huggingface.co./BitStarWalkin/SuperCorrect-7B/ | Name | Quant method | Size | | ---- | ---- | ---- | | [SuperCorrect-7B.Q2_K.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q2_K.gguf) | Q2_K | 2.81GB | | [SuperCorrect-7B.Q3_K_S.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q3_K_S.gguf) | Q3_K_S | 3.25GB | | [SuperCorrect-7B.Q3_K.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q3_K.gguf) | Q3_K | 3.55GB | | [SuperCorrect-7B.Q3_K_M.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q3_K_M.gguf) | Q3_K_M | 3.55GB | | [SuperCorrect-7B.Q3_K_L.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q3_K_L.gguf) | Q3_K_L | 3.81GB | | [SuperCorrect-7B.IQ4_XS.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.IQ4_XS.gguf) | IQ4_XS | 3.96GB | | [SuperCorrect-7B.Q4_0.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q4_0.gguf) | Q4_0 | 4.13GB | | [SuperCorrect-7B.IQ4_NL.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.IQ4_NL.gguf) | IQ4_NL | 4.16GB | | [SuperCorrect-7B.Q4_K_S.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q4_K_S.gguf) | Q4_K_S | 4.15GB | | [SuperCorrect-7B.Q4_K.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q4_K.gguf) | Q4_K | 4.36GB | | [SuperCorrect-7B.Q4_K_M.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q4_K_M.gguf) | Q4_K_M | 4.36GB | | [SuperCorrect-7B.Q4_1.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q4_1.gguf) | Q4_1 | 4.54GB | | [SuperCorrect-7B.Q5_0.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q5_0.gguf) | Q5_0 | 4.95GB | | [SuperCorrect-7B.Q5_K_S.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q5_K_S.gguf) | Q5_K_S | 4.95GB | | [SuperCorrect-7B.Q5_K.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q5_K.gguf) | Q5_K | 5.07GB | | [SuperCorrect-7B.Q5_K_M.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q5_K_M.gguf) | Q5_K_M | 5.07GB | | [SuperCorrect-7B.Q5_1.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q5_1.gguf) | Q5_1 | 5.36GB | | [SuperCorrect-7B.Q6_K.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q6_K.gguf) | Q6_K | 5.82GB | | [SuperCorrect-7B.Q8_0.gguf](https://huggingface.co./RichardErkhov/BitStarWalkin_-_SuperCorrect-7B-gguf/blob/main/SuperCorrect-7B.Q8_0.gguf) | Q8_0 | 7.54GB | Original model description: --- license: apache-2.0 language: - en metrics: - accuracy base_model: - Qwen/Qwen2.5-Math-7B-Instruct library_name: transformers pipeline_tag: question-answering --- ## SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights > [SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights](https://arxiv.org/abs/2410.09008) > [Ling Yang\*](https://yangling0818.github.io/), [Zhaochen Yu*](https://github.com/BitCodingWalkin), [Tianjun Zhang](https://tianjunz.github.io/), [Minkai Xu](https://minkaixu.com/), [Joseph E. Gonzalez](https://people.eecs.berkeley.edu/~jegonzal/),[Bin Cui](https://cuibinpku.github.io/), [Shuicheng Yan](https://yanshuicheng.info/) > > Peking University, Skywork AI, UC Berkeley, Stanford University
## Introduction  This repo provides the official implementation of **SuperCorrect** a novel two-stage fine-tuning method for improving both reasoning accuracy and self-correction ability for LLMs. Notably, our **SupperCorrect-7B** model significantly surpasses powerful **DeepSeekMath-7B by 7.8%/5.3% and Qwen2.5-Math-7B by 15.1%/6.3% on MATH/GSM8K benchmarks**, achieving new SOTA performance among all 7B models.