Add pipeline tag, library name, and datasets
#1
by
nielsr
HF staff
- opened
README.md
CHANGED
@@ -1,3 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
Quantization made by Richard Erkhov.
|
2 |
|
3 |
[Github](https://github.com/RichardErkhov)
|
@@ -47,10 +62,11 @@ metrics:
|
|
47 |
base_model:
|
48 |
- Qwen/Qwen2.5-Math-7B-Instruct
|
49 |
library_name: transformers
|
|
|
50 |
---
|
51 |
## SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights
|
52 |
|
53 |
-
> [SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights](
|
54 |
> [Ling Yang\*](https://yangling0818.github.io/), [Zhaochen Yu*](https://github.com/BitCodingWalkin), [Tianjun Zhang](https://tianjunz.github.io/), [Minkai Xu](https://minkaixu.com/), [Joseph E. Gonzalez](https://people.eecs.berkeley.edu/~jegonzal/),[Bin Cui](https://cuibinpku.github.io/), [Shuicheng Yan](https://yanshuicheng.info/)
|
55 |
>
|
56 |
> Peking University, Skywork AI, UC Berkeley, Stanford University
|
@@ -152,21 +168,21 @@ We evaluate our SupperCorrect-7B on two widely used English math benchmarks GSM8
|
|
152 |
## Citation
|
153 |
|
154 |
```bash
|
155 |
-
@
|
156 |
-
title={SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights}
|
157 |
author={Yang, Ling and Yu, Zhaochen and Zhang, Tianjun and Xu, Minkai and Gonzalez, Joseph E and Cui, Bin and Yan, Shuicheng},
|
158 |
-
|
159 |
-
year={
|
160 |
}
|
|
|
161 |
@article{yang2024buffer,
|
162 |
title={Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models},
|
163 |
author={Yang, Ling and Yu, Zhaochen and Zhang, Tianjun and Cao, Shiyi and Xu, Minkai and Zhang, Wentao and Gonzalez, Joseph E and Cui, Bin},
|
164 |
-
journal={
|
165 |
year={2024}
|
166 |
}
|
167 |
```
|
168 |
|
169 |
## Acknowledgements
|
170 |
|
171 |
-
Our SuperCorrect is a two-stage fine-tuning model which based on several extraordinary open-source models like [Qwen2.5-Math](https://github.com/QwenLM/Qwen2.5-Math), [DeepSeek-Math](https://github.com/deepseek-ai/DeepSeek-Math), [Llama3-Series](https://github.com/meta-llama/llama3). Our evaluation method is based on the code base of outstanding works like [Qwen2.5-Math](https://github.com/QwenLM/Qwen2.5-Math) and [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness). We also want to express our gratitude for amazing works such as [BoT](https://github.com/YangLing0818/buffer-of-thought-llm) which provides the idea of thought template.
|
172 |
-
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
metrics:
|
6 |
+
- accuracy
|
7 |
+
base_model:
|
8 |
+
- Qwen/Qwen2.5-Math-7B-Instruct
|
9 |
+
library_name: transformers
|
10 |
+
pipeline_tag: question-answering
|
11 |
+
datasets:
|
12 |
+
- MATH
|
13 |
+
- GSM8K
|
14 |
+
---
|
15 |
+
|
16 |
Quantization made by Richard Erkhov.
|
17 |
|
18 |
[Github](https://github.com/RichardErkhov)
|
|
|
62 |
base_model:
|
63 |
- Qwen/Qwen2.5-Math-7B-Instruct
|
64 |
library_name: transformers
|
65 |
+
pipeline_tag: question-answering
|
66 |
---
|
67 |
## SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights
|
68 |
|
69 |
+
> [SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights](https://arxiv.org/abs/2410.09008)
|
70 |
> [Ling Yang\*](https://yangling0818.github.io/), [Zhaochen Yu*](https://github.com/BitCodingWalkin), [Tianjun Zhang](https://tianjunz.github.io/), [Minkai Xu](https://minkaixu.com/), [Joseph E. Gonzalez](https://people.eecs.berkeley.edu/~jegonzal/),[Bin Cui](https://cuibinpku.github.io/), [Shuicheng Yan](https://yanshuicheng.info/)
|
71 |
>
|
72 |
> Peking University, Skywork AI, UC Berkeley, Stanford University
|
|
|
168 |
## Citation
|
169 |
|
170 |
```bash
|
171 |
+
@inproceedings{yang2025supercorrect,
|
172 |
+
title={SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights},
|
173 |
author={Yang, Ling and Yu, Zhaochen and Zhang, Tianjun and Xu, Minkai and Gonzalez, Joseph E and Cui, Bin and Yan, Shuicheng},
|
174 |
+
booktitle={International Conference on Learning Representations},
|
175 |
+
year={2025}
|
176 |
}
|
177 |
+
|
178 |
@article{yang2024buffer,
|
179 |
title={Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models},
|
180 |
author={Yang, Ling and Yu, Zhaochen and Zhang, Tianjun and Cao, Shiyi and Xu, Minkai and Zhang, Wentao and Gonzalez, Joseph E and Cui, Bin},
|
181 |
+
journal={Advances in Neural Information Processing Systems},
|
182 |
year={2024}
|
183 |
}
|
184 |
```
|
185 |
|
186 |
## Acknowledgements
|
187 |
|
188 |
+
Our SuperCorrect is a two-stage fine-tuning model which based on several extraordinary open-source models like [Qwen2.5-Math](https://github.com/QwenLM/Qwen2.5-Math), [DeepSeek-Math](https://github.com/deepseek-ai/DeepSeek-Math), [Llama3-Series](https://github.com/meta-llama/llama3). Our evaluation method is based on the code base of outstanding works like [Qwen2.5-Math](https://github.com/QwenLM/Qwen2.5-Math) and [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness). We also want to express our gratitude for amazing works such as [BoT](https://github.com/YangLing0818/buffer-of-thought-llm) which provides the idea of thought template.
|
|