jmeadows17 commited on
Commit
85265ae
1 Parent(s): b9e46de

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -5,11 +5,11 @@ library_name: peft
5
 
6
  **Overview**
7
 
8
- DerivationGeneration8B is a QLoRA fine-tuned from a quantised LLaMa-3.1-8B checkpoint on 15K (LaTeX) synthetic mathematical derivations (containing 4 - 10 equations) via a custom early stopping script using ROUGE as the validation metric. This LoRA was trained for 6 epochs, and the approach outperforms MathT5 ```https://huggingface.co/jmeadows17/MathT5-large``` in both in-distribution and perturbed evaluation cases presented in related work ```https://arxiv.org/abs/2307.09998```.
9
 
10
  **How to use**
11
 
12
- A notebook for inference is available here ```https://github.com/jmeadows17/deriving-equations-with-LLMs/blob/main/llama_evaluation.ipynb```. Training scripts are also available in the repository.
13
 
14
  **Example prompt**
15
 
 
5
 
6
  **Overview**
7
 
8
+ DerivationGeneration8B is a QLoRA fine-tuned from a quantised LLaMa-3.1-8B checkpoint on 15K (LaTeX) synthetic mathematical derivations (containing 4 - 10 equations) via a custom early stopping script using ROUGE as the validation metric (total 6 epochs). This approach outperforms [MathT5](https://huggingface.co/jmeadows17/MathT5-large) in both in-distribution and perturbed evaluation cases presented in related work ```https://arxiv.org/abs/2307.09998```.
9
 
10
  **How to use**
11
 
12
+ A notebook for inference is available [here](https://github.com/jmeadows17/deriving-equations-with-LLMs/blob/main/llama_evaluation.ipynb). Training scripts are also available in the repository.
13
 
14
  **Example prompt**
15