nikhil07prakash commited on
Commit
5f4fe07
·
verified ·
1 Parent(s): 939aa50

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -6,7 +6,7 @@ license: mit
6
 
7
  <!-- Provide a quick summary of what the model is/does. -->
8
 
9
- This model is a vanilla fine-tuned version of the [Llama-7B](https://huggingface.co/huggyllama/llama-7b) model on synthetically generated arithmetic tasks. It was introduced in [this](https://openreview.net/forum?id=8sKcAWOf2D) paper. It is very similar to [Goat-7B](https://github.com/liutiedong/goat), except it was trained without LoRA.
10
 
11
  For inquiries about checkpoints during the fine-tuning process, kindly reach out to [Nikhil](mailto:[email protected]) via email.
12
 
@@ -19,14 +19,14 @@ For inquiries about checkpoints during the fine-tuning process, kindly reach out
19
  - **Developed by:** [Nikhil Prakash](https://nix07.github.io/)
20
  - **Model type:** Autoregressive Decoder-only Language Model
21
  - **License:** MIT License
22
- - **Finetuned from model [optional]:** [Llama-7B](https://huggingface.co/huggyllama/llama-7b)
23
 
24
- ### Model Sources [optional]
25
 
26
  <!-- Provide the basic links for the model. -->
27
 
28
- - **Repository:** TODO
29
- - **Paper [optional]:** [Fine-Tuning Enhances Existing Mechanisms: A Case Study on Entity Tracking](https://openreview.net/forum?id=8sKcAWOf2D)
30
 
31
  ## How to Get Started with the Model
32
 
@@ -37,7 +37,7 @@ from transformers import AutoModel
37
  model = AutoModel.from_pretrained("nikhil07prakash/float-7b")
38
  ```
39
 
40
- ## Citation [optional]
41
 
42
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
43
 
 
6
 
7
  <!-- Provide a quick summary of what the model is/does. -->
8
 
9
+ This model is a fully fine-tuned version of the [Llama-7B](https://huggingface.co/huggyllama/llama-7b) model on synthetically generated arithmetic tasks. It was introduced in [this](https://openreview.net/forum?id=8sKcAWOf2D) paper. It is very similar to [Goat-7B](https://github.com/liutiedong/goat), except it was trained without LoRA.
10
 
11
  For inquiries about checkpoints during the fine-tuning process, kindly reach out to [Nikhil](mailto:[email protected]) via email.
12
 
 
19
  - **Developed by:** [Nikhil Prakash](https://nix07.github.io/)
20
  - **Model type:** Autoregressive Decoder-only Language Model
21
  - **License:** MIT License
22
+ - **Finetuned from model:** [Llama-7B](https://huggingface.co/huggyllama/llama-7b)
23
 
24
+ ### Model Sources
25
 
26
  <!-- Provide the basic links for the model. -->
27
 
28
+ - **Repository:** [Link](https://github.com/Nix07/finetuning/)
29
+ - **Paper :** [Fine-Tuning Enhances Existing Mechanisms: A Case Study on Entity Tracking](https://openreview.net/forum?id=8sKcAWOf2D)
30
 
31
  ## How to Get Started with the Model
32
 
 
37
  model = AutoModel.from_pretrained("nikhil07prakash/float-7b")
38
  ```
39
 
40
+ ## Citation
41
 
42
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
43