prompts for humaneval

#50
by kcdharma - opened

Hi,
In the paper I see that StarCoder-Prompted is better than StarCoder on HumanEval dataset. Could you please let me know where can I find these prompt examples? Thank you.

BigCode org

The prompt is mentioned in the paper section 6.1.1. We basically added this prefix to the HumanEval prompts <filename>solutions/solution_1.py\n# Here is the correct implementation of the code exercise\n before feeding them to the model.

loubnabnl changed discussion status to closed

Thank you @loubnabnl . But, where can I find the solution1.py? and does the code internally loads the contents of this solution and adds that as a prefix to the original prompt?

BigCode org

That's actually just text that we add at the beginning of each problem since we conditionned on file paths during pre-training. So the model tends to give better completions when we indicate that the code comes from a file with the path solutions/solution_1.py

Well, this is interesting, it's only that " solutions/solution_1.py\n# Here is the correct implementation of the code exercise\n" as the prefix and that increases the accuracy of the model from 34% to 40%. It feels quite counterintuitive. I feel like the model should produce the same result with/without this input (if the model understands something about the question). did you guys further explore this direction and found out something why you get these results? Thank you.

Sign up or log in to comment