nielsr HF staff commited on
Commit
5733aa5
verified
1 Parent(s): 1125c72

Add pipeline tag, library name, and datasets

Browse files

This PR adds the appropriate `pipeline_tag` (question-answering) and library name (transformers) to the model card. Also adds datasets MATH and GSM8K to the metadata.

Files changed (1) hide show
  1. README.md +24 -8
README.md CHANGED
@@ -1,3 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  Quantization made by Richard Erkhov.
2
 
3
  [Github](https://github.com/RichardErkhov)
@@ -47,10 +62,11 @@ metrics:
47
  base_model:
48
  - Qwen/Qwen2.5-Math-7B-Instruct
49
  library_name: transformers
 
50
  ---
51
  ## SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights
52
 
53
- > [SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights](link)
54
  > [Ling Yang\*](https://yangling0818.github.io/), [Zhaochen Yu*](https://github.com/BitCodingWalkin), [Tianjun Zhang](https://tianjunz.github.io/), [Minkai Xu](https://minkaixu.com/), [Joseph E. Gonzalez](https://people.eecs.berkeley.edu/~jegonzal/),[Bin Cui](https://cuibinpku.github.io/), [Shuicheng Yan](https://yanshuicheng.info/)
55
  >
56
  > Peking University, Skywork AI, UC Berkeley, Stanford University
@@ -152,21 +168,21 @@ We evaluate our SupperCorrect-7B on two widely used English math benchmarks GSM8
152
  ## Citation
153
 
154
  ```bash
155
- @article{yang2024supercorrect,
156
- title={SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights}
157
  author={Yang, Ling and Yu, Zhaochen and Zhang, Tianjun and Xu, Minkai and Gonzalez, Joseph E and Cui, Bin and Yan, Shuicheng},
158
- journal={arXiv preprint arXiv:2410.09008},
159
- year={2024}
160
  }
 
161
  @article{yang2024buffer,
162
  title={Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models},
163
  author={Yang, Ling and Yu, Zhaochen and Zhang, Tianjun and Cao, Shiyi and Xu, Minkai and Zhang, Wentao and Gonzalez, Joseph E and Cui, Bin},
164
- journal={arXiv preprint arXiv:2406.04271},
165
  year={2024}
166
  }
167
  ```
168
 
169
  ## Acknowledgements
170
 
171
- Our SuperCorrect is a two-stage fine-tuning model which based on several extraordinary open-source models like [Qwen2.5-Math](https://github.com/QwenLM/Qwen2.5-Math), [DeepSeek-Math](https://github.com/deepseek-ai/DeepSeek-Math), [Llama3-Series](https://github.com/meta-llama/llama3). Our evaluation method is based on the code base of outstanding works like [Qwen2.5-Math](https://github.com/QwenLM/Qwen2.5-Math) and [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness). We also want to express our gratitude for amazing works such as [BoT](https://github.com/YangLing0818/buffer-of-thought-llm) which provides the idea of thought template.
172
-
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ metrics:
6
+ - accuracy
7
+ base_model:
8
+ - Qwen/Qwen2.5-Math-7B-Instruct
9
+ library_name: transformers
10
+ pipeline_tag: question-answering
11
+ datasets:
12
+ - MATH
13
+ - GSM8K
14
+ ---
15
+
16
  Quantization made by Richard Erkhov.
17
 
18
  [Github](https://github.com/RichardErkhov)
 
62
  base_model:
63
  - Qwen/Qwen2.5-Math-7B-Instruct
64
  library_name: transformers
65
+ pipeline_tag: question-answering
66
  ---
67
  ## SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights
68
 
69
+ > [SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights](https://arxiv.org/abs/2410.09008)
70
  > [Ling Yang\*](https://yangling0818.github.io/), [Zhaochen Yu*](https://github.com/BitCodingWalkin), [Tianjun Zhang](https://tianjunz.github.io/), [Minkai Xu](https://minkaixu.com/), [Joseph E. Gonzalez](https://people.eecs.berkeley.edu/~jegonzal/),[Bin Cui](https://cuibinpku.github.io/), [Shuicheng Yan](https://yanshuicheng.info/)
71
  >
72
  > Peking University, Skywork AI, UC Berkeley, Stanford University
 
168
  ## Citation
169
 
170
  ```bash
171
+ @inproceedings{yang2025supercorrect,
172
+ title={SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights},
173
  author={Yang, Ling and Yu, Zhaochen and Zhang, Tianjun and Xu, Minkai and Gonzalez, Joseph E and Cui, Bin and Yan, Shuicheng},
174
+ booktitle={International Conference on Learning Representations},
175
+ year={2025}
176
  }
177
+
178
  @article{yang2024buffer,
179
  title={Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models},
180
  author={Yang, Ling and Yu, Zhaochen and Zhang, Tianjun and Cao, Shiyi and Xu, Minkai and Zhang, Wentao and Gonzalez, Joseph E and Cui, Bin},
181
+ journal={Advances in Neural Information Processing Systems},
182
  year={2024}
183
  }
184
  ```
185
 
186
  ## Acknowledgements
187
 
188
+ Our SuperCorrect is a two-stage fine-tuning model which based on several extraordinary open-source models like [Qwen2.5-Math](https://github.com/QwenLM/Qwen2.5-Math), [DeepSeek-Math](https://github.com/deepseek-ai/DeepSeek-Math), [Llama3-Series](https://github.com/meta-llama/llama3). Our evaluation method is based on the code base of outstanding works like [Qwen2.5-Math](https://github.com/QwenLM/Qwen2.5-Math) and [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness). We also want to express our gratitude for amazing works such as [BoT](https://github.com/YangLing0818/buffer-of-thought-llm) which provides the idea of thought template.