|
model: |
|
checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors" |
|
pretrained_model_name_or_path: fal/AuraFlow-v0.3 |
|
|
|
dtype: bfloat16 |
|
|
|
denoiser: |
|
use_flash_attn: true |
|
|
|
use_rope: True |
|
rope_theta: 10000 |
|
rope_dim_sizes: [32, 112, 112] |
|
|
|
peft: |
|
type: lora |
|
rank: 4 |
|
alpha: 1.0 |
|
dropout: 0.0 |
|
|
|
dtype: bfloat16 |
|
|
|
include_keys: [".attn."] |
|
exclude_keys: ["text_encoder", "vae", "t_embedder", "final_linear"] |
|
|
|
dataset: |
|
folder: "data/pexels-1k-random" |
|
num_repeats: 2 |
|
batch_size: 2 |
|
|
|
bucket_base_size: 1024 |
|
step: 128 |
|
min_size: 384 |
|
do_upscale: false |
|
|
|
caption_processors: [] |
|
|
|
optimizer: |
|
name: "schedulefree.RAdamScheduleFree" |
|
|
|
args: |
|
lr: 0.008 |
|
|
|
scheduler: |
|
|
|
|
|
|
|
tracker: |
|
project_name: "auraflow-rope-1" |
|
loggers: |
|
- wandb |
|
|
|
saving: |
|
strategy: |
|
per_epochs: 1 |
|
per_steps: null |
|
save_last: true |
|
|
|
callbacks: |
|
- type: "hf_hub" |
|
name: "rope-1" |
|
save_dir: "./output/rope-3" |
|
|
|
hub_id: "p1atdev/afv03-lora" |
|
dir_in_repo: "rope-3" |
|
|
|
seed: 42 |
|
num_train_epochs: 5 |
|
|
|
trainer: |
|
|
|
|
|
gradient_checkpointing: true |
|
|
|
torch_compile: true |
|
torch_compile_args: |
|
mode: max-autotune |
|
fullgraph: true |
|
fp32_matmul_precision: "medium" |
|
|