ayanahye commited on
Commit
9732fa4
1 Parent(s): 64bc8e5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -6
README.md CHANGED
@@ -7,6 +7,8 @@ tags:
7
  model-index:
8
  - name: DocGPT-ft
9
  results: []
 
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -14,21 +16,20 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # DocGPT-ft
16
 
17
- This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 2.3387
20
 
21
  ## Model description
22
 
23
- More information needed
24
 
25
  ## Intended uses & limitations
26
 
27
- More information needed
28
 
29
  ## Training and evaluation data
30
 
31
- More information needed
32
 
33
  ## Training procedure
34
 
 
7
  model-index:
8
  - name: DocGPT-ft
9
  results: []
10
+ datasets:
11
+ - lavita/ChatDoctor-HealthCareMagic-100k
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
16
 
17
  # DocGPT-ft
18
 
19
+ This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on the lavita/ChatDoctor-HealthCareMagic-100k dataset.
20
+
 
21
 
22
  ## Model description
23
 
24
+ Uses parameter efficient fine-tuning for QLora
25
 
26
  ## Intended uses & limitations
27
 
28
+ The intended use is just for fun.
29
 
30
  ## Training and evaluation data
31
 
32
+ The training set was 90% of the data and testing set was 10%. Only a small percentage of the data was used to reduce training time.
33
 
34
  ## Training procedure
35