p1atdev commited on
Commit
1180a89
·
verified ·
1 Parent(s): ec7428b

Upload rope_migration.yml

Browse files
Files changed (1) hide show
  1. rope-2/rope_migration.yml +79 -0
rope-2/rope_migration.yml ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors"
3
+ pretrained_model_name_or_path: fal/AuraFlow-v0.3
4
+
5
+ dtype: bfloat16
6
+
7
+ denoiser:
8
+ use_flash_attn: true
9
+
10
+ use_rope: True
11
+ rope_theta: 10000
12
+ rope_dim_sizes: [32, 112, 112]
13
+
14
+ peft:
15
+ type: lora
16
+ rank: 16
17
+ alpha: 1.0
18
+ dropout: 0.0
19
+
20
+ dtype: bfloat16
21
+
22
+ # include_keys: [".attn."]
23
+ include_keys: [".attn.", ".mlp.", ".modC.", ".modC.", ".modX."]
24
+ exclude_keys: ["text_encoder", "vae", "t_embedder", "final_linear"]
25
+
26
+ dataset:
27
+ folder: "data/pexels-1k-random"
28
+ num_repeats: 2
29
+ batch_size: 2
30
+
31
+ bucket_base_size: 1024
32
+ step: 128
33
+ min_size: 384
34
+ do_upscale: false
35
+
36
+ caption_processors: []
37
+
38
+ optimizer:
39
+ name: "schedulefree.RAdamScheduleFree"
40
+ # name: "bitsandbytes.optim.AdamW8bit"
41
+ args:
42
+ lr: 0.001
43
+
44
+ scheduler:
45
+ # name: "torch.optim.lr_scheduler.ConstantLR"
46
+ # args: {}
47
+
48
+ tracker:
49
+ project_name: "auraflow-rope-1"
50
+ loggers:
51
+ - wandb
52
+
53
+ saving:
54
+ strategy:
55
+ per_epochs: 1
56
+ per_steps: null
57
+ save_last: true
58
+
59
+ callbacks:
60
+ - type: "hf_hub" # or "hf_hub" to push to hub
61
+ name: "rope-1"
62
+ save_dir: "./output/rope-2"
63
+
64
+ hub_id: "p1atdev/afv03-lora"
65
+ dir_in_repo: "rope-2"
66
+
67
+ seed: 42
68
+ num_train_epochs: 20
69
+
70
+ trainer:
71
+ # debug_mode: "1step"
72
+
73
+ gradient_checkpointing: true
74
+
75
+ torch_compile: true
76
+ torch_compile_args:
77
+ mode: max-autotune
78
+ fullgraph: true
79
+ fp32_matmul_precision: "medium"