prsdm commited on
Commit
66fac74
1 Parent(s): 9536c8b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -0
README.md CHANGED
@@ -11,3 +11,32 @@ tags:
11
  - medical
12
  ---
13
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  - medical
12
  ---
13
 
14
+
15
+ Updated on 14th August 2024
16
+ # Fine-tuning Large Language Models
17
+ <p>
18
+ 😸<a href="https://github.com/prsdm/fine-tuning-llms/blob/main/Fine-tuning-phi-2-model.ipynb">GitHub</a> •📝<a href="https://medium.com/@prasadmahamulkar/fine-tuning-phi-2-a-step-by-step-guide-e672e7f1d009">Article</a> • Models & Datasets on: 🤗<a href="https://huggingface.co/prsdm">Hugging Face</a>
19
+ </p>
20
+
21
+ This repository provides a collection of Jupyter notebooks that demonstrate how to fine-tune large language models using various tools and techniques.
22
+
23
+ fine-tuning or instruction tuning is the process where the pre-trained model is further trained on the smaller dataset to adapt its knowledge for a specific task or domain. This process tweaks the model’s parameters to perform specific tasks. In fine-tuning, there are two methods:
24
+
25
+ ### Supervised fine tuning (SFT):
26
+ In SFT, the model is trained on a labeled dataset. The labeled dataset typically contains examples of instruction (input) and response (output) pairs relevant to the task. In this process, the model learns how to respond to specific instructions.
27
+
28
+ ### Reinforcement Learning from Human Feedback (RLHF):
29
+ In RLHF, the model interacts with users, generates responses, and receives feedback in the form of reinforcement signals. Basically, the model learns and improves its performance based on the feedback it receives.
30
+
31
+ ### Details of each notebook:
32
+
33
+ | LLMs | Description | Dataset | Notebooks | Tuned Model |
34
+ |---------------------------|-------------|---------|-----------|-------------|
35
+ | [Phi-2](https://huggingface.co/microsoft/phi-2) | This model has been fine-tuned on a Medical dataset to answer questions related to diseases and symptoms. (used SFT method) | [Dataset](https://huggingface.co/datasets/prsdm/MedQuad-phi2-1k) | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/prsdm/fine-tuning-llms/blob/main/Fine-tuning-phi-2-model.ipynb) | [phi-2-medquad](https://huggingface.co/prsdm/phi-2-medquad) |
36
+ | [llama-2](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) | This model has been fine-tuned on a dataset with human-generated prompts to answer questions related to general knowledge. (used SFT method) | [Dataset](https://huggingface.co/datasets/prsdm/finance-llama2-1k) | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/prsdm/fine-tuning-llms/blob/main/Fine-tuning-llama-2-model.ipynb) | [llama-2-finance](https://huggingface.co/prsdm/llama-2-finance) |
37
+
38
+
39
+ ### Diagram:
40
+
41
+ ![diagram](https://github.com/user-attachments/assets/b84531b3-9935-4e2f-bd05-e0f88f95edb6)
42
+