shirsh10mall commited on
Commit
b675dea
1 Parent(s): fdd0aa1

NumExample_75000_Epoch_2

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. adapter_config.json +2 -2
  3. adapter_model.bin +1 -1
README.md CHANGED
@@ -18,4 +18,4 @@ The following `bitsandbytes` quantization config was used during training:
18
  ### Framework versions
19
 
20
 
21
- - PEFT 0.5.0.dev0
 
18
  ### Framework versions
19
 
20
 
21
+ - PEFT 0.6.0.dev0
adapter_config.json CHANGED
@@ -7,8 +7,8 @@
7
  "init_lora_weights": true,
8
  "layers_pattern": null,
9
  "layers_to_transform": null,
10
- "lora_alpha": 16,
11
- "lora_dropout": 0.05,
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
  "r": 8,
 
7
  "init_lora_weights": true,
8
  "layers_pattern": null,
9
  "layers_to_transform": null,
10
+ "lora_alpha": 512,
11
+ "lora_dropout": 0.01,
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
  "r": 8,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c02b673cb84b2cae7784a9fd3aafcd7999ff6917f73295ff0696ec5ba6dcb9b
3
  size 9543245
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5717e6ceb65601c11660ffde54a9a336d2190ef9b1b4f4f92be505bd7f3ffed9
3
  size 9543245