p1atdev commited on
Commit
39f87ea
·
verified ·
1 Parent(s): 680bc11

Upload rope_migration.yml

Browse files
Files changed (1) hide show
  1. rope-11/rope_migration.yml +83 -0
rope-11/rope_migration.yml ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors"
3
+ pretrained_model_name_or_path: fal/AuraFlow-v0.3
4
+
5
+ dtype: bfloat16
6
+
7
+ denoiser:
8
+ use_flash_attn: true
9
+
10
+ use_rope: True
11
+ rope_theta: 10000
12
+ rope_dim_sizes: [32, 112, 112]
13
+
14
+ noise_prediction_loss: true
15
+ migration_loss: true
16
+ # prior_preservation_loss: true
17
+
18
+ peft:
19
+ type: lora
20
+ rank: 4
21
+ alpha: 1.0
22
+ dropout: 0.0
23
+
24
+ dtype: bfloat16
25
+
26
+ # include_keys: [".attn.", ".mlpC.", ".mlpX.", ".mlp.", ".modC.", ".modCX.", ".modX."]
27
+ include_keys: [".attn.", ".mlpC.", ".mlpX.", ".mlp."]
28
+ exclude_keys: ["text_encoder", "vae", "t_embedder", "final_linear"]
29
+
30
+ dataset:
31
+ folder: "data/pexels-1k-random"
32
+ num_repeats: 2
33
+ batch_size: 2
34
+
35
+ bucket_base_size: 1024
36
+ step: 128
37
+ min_size: 384
38
+ do_upscale: false
39
+
40
+ caption_processors: []
41
+
42
+ optimizer:
43
+ name: "schedulefree.RAdamScheduleFree"
44
+ # name: "bitsandbytes.optim.AdamW8bit"
45
+ args:
46
+ lr: 0.005
47
+
48
+ scheduler:
49
+ # name: "torch.optim.lr_scheduler.ConstantLR"
50
+ # args: {}
51
+
52
+ tracker:
53
+ project_name: "auraflow-rope-1"
54
+ loggers:
55
+ - wandb
56
+
57
+ saving:
58
+ strategy:
59
+ per_epochs: 1
60
+ per_steps: null
61
+ save_last: true
62
+
63
+ callbacks:
64
+ - type: "hf_hub" # or "hf_hub" to push to hub
65
+ name: "rope-11"
66
+ save_dir: "./output/rope-11"
67
+
68
+ hub_id: "p1atdev/afv03-lora"
69
+ dir_in_repo: "rope-11"
70
+
71
+ seed: 42
72
+ num_train_epochs: 20
73
+
74
+ trainer:
75
+ # debug_mode: "1step"
76
+
77
+ gradient_checkpointing: true
78
+
79
+ torch_compile: true
80
+ torch_compile_args:
81
+ mode: max-autotune
82
+ fullgraph: true
83
+ fp32_matmul_precision: "medium"