eswardivi commited on
Commit
4356dc0
1 Parent(s): 853f55b

Upload 4 files

Browse files
Files changed (4) hide show
  1. README.md +20 -0
  2. adapter_config.json +22 -0
  3. adapter_model.bin +3 -0
  4. config.md +62 -0
README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: True
9
+ - load_in_4bit: False
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: fp4
15
+ - bnb_4bit_use_double_quant: False
16
+ - bnb_4bit_compute_dtype: float32
17
+ ### Framework versions
18
+
19
+
20
+ - PEFT 0.4.0.dev0
adapter_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "yahma/llama-7b-hf",
3
+ "bias": "none",
4
+ "fan_in_fan_out": false,
5
+ "inference_mode": true,
6
+ "init_lora_weights": true,
7
+ "layers_pattern": null,
8
+ "layers_to_transform": null,
9
+ "lora_alpha": 16,
10
+ "lora_dropout": 0.05,
11
+ "modules_to_save": null,
12
+ "peft_type": "LORA",
13
+ "r": 16,
14
+ "revision": null,
15
+ "target_modules": [
16
+ "q_proj",
17
+ "k_proj",
18
+ "v_proj",
19
+ "o_proj"
20
+ ],
21
+ "task_type": "CAUSAL_LM"
22
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8d2f68cfad1b8053786403f759e7ecac6396803886afee7172135b296ab1370
3
+ size 67201357
config.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Trained using https://github.com/tloen/alpaca-lora
2
+ with removing the lines
3
+
4
+ ```
5
+ old_state_dict = model.state_dict
6
+ model.state_dict = (
7
+ lambda self, *_, **__: get_peft_model_state_dict(
8
+ self, old_state_dict()
9
+ )
10
+ ).__get__(model, type(model))
11
+ ```
12
+
13
+ causing problem.
14
+
15
+ base_model: yahma/llama-7b-hf
16
+ data_path: prognosis/medical_qa_alpaca
17
+ output_dir: ./lora-alpaca
18
+ batch_size: 128
19
+ micro_batch_size: 8
20
+ num_epochs: 5
21
+ learning_rate: 0.0003
22
+ cutoff_len: 512
23
+ val_set_size: 0.1
24
+ lora_r: 16
25
+ lora_alpha: 16
26
+ lora_dropout: 0.05
27
+ lora_target_modules: ['q_proj', 'k_proj', 'v_proj', 'o_proj']
28
+ train_on_inputs: True
29
+ add_eos_token: False
30
+ group_by_length: True
31
+ wandb_project: medical_alpaca_hf
32
+ wandb_run_name: run_3
33
+ wandb_watch:
34
+ wandb_log_model:
35
+ resume_from_checkpoint: False
36
+ prompt template: alpaca
37
+
38
+
39
+ ### Command used
40
+
41
+
42
+ Finetuning
43
+ ```
44
+ python finetune.py --base_model 'yahma/llama-7b-hf' --data_path 'prognosis/medical_qa_alpaca' --output_dir './lora-alpaca' --wandb_project 'medical_alpaca_hf' --wandb_run_name 'run_3' --lora_target_modules '[q_proj,k_proj,v_proj,o_proj]' --num_epochs 5 --cutoff_len 512 --group_by_length --val_set_size 0.1 --lora_r=16 --micro_batch_size=8
45
+ ```
46
+
47
+ Generating
48
+
49
+ ```
50
+ python generate.py \
51
+ --load_8bit \
52
+ --base_model 'yahma/llama-7b-hf' \
53
+ --lora_weights 'alpaca-lora/lora-alpaca' \
54
+ --share_gradio
55
+ ```
56
+
57
+ git lfs
58
+
59
+ ```
60
+ curl -s https://packagecloud.io/install/repositories/github/git-lfs/script.deb.sh | sudo bash
61
+ sudo apt-get install git-lfs
62
+ ```