raman07 commited on
Commit
1686a18
·
verified ·
1 Parent(s): ce4e283

End of training

Browse files
README.md ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ license: creativeml-openrail-m
4
+ base_model: raman07/pixart-alpha-256x256
5
+ tags:
6
+ - pixart-alpha
7
+ - medical-pixart-alpha
8
+ - text-to-image
9
+ - diffusers
10
+ - lora
11
+ inference: true
12
+ ---
13
+
14
+ # LoRA text2image fine-tuning - raman07/LR_1e-05
15
+ These are LoRA adaption weights for raman07/pixart-alpha-256x256. The weights were fine-tuned on the MIMIC-CXR dataset.
adapter_config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": {
4
+ "base_model_class": "Transformer2DModel",
5
+ "parent_library": "diffusers.models.transformers.transformer_2d"
6
+ },
7
+ "base_model_name_or_path": null,
8
+ "bias": "none",
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": "gaussian",
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "lora_alpha": 8,
15
+ "lora_dropout": 0.0,
16
+ "modules_to_save": null,
17
+ "peft_type": "LORA",
18
+ "r": 4,
19
+ "rank_pattern": {},
20
+ "revision": null,
21
+ "target_modules": [
22
+ "linear_2",
23
+ "ff.net.2",
24
+ "to_v",
25
+ "ff.net.0.proj",
26
+ "proj",
27
+ "to_k",
28
+ "to_out.0",
29
+ "linear",
30
+ "to_q",
31
+ "proj_out",
32
+ "linear_1",
33
+ "proj_in"
34
+ ],
35
+ "task_type": null
36
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4379a2597f9cae7dd9fda2b261e709992d4df74fdcbf8a539e940fb4d52e21ea
3
+ size 13973122
transformer_lora/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca75af625f66c008f7944ceb7a2fc25b2c7dae92e6ea0f75a8d4296d20cc9888
3
+ size 13844488