model: checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors" pretrained_model_name_or_path: fal/AuraFlow-v0.3 dtype: bfloat16 denoiser: use_flash_attn: true peft: type: lora rank: 16 alpha: 1.0 dropout: 0.0 dtype: bfloat16 include_keys: [".attn.", ".mlp.", ".modC.", ".modC.", ".modX."] exclude_keys: ["text_encoder", "vae", "t_embedder", "final_linear"] dataset: folder: "data/pvc" num_repeats: 2 batch_size: 2 bucket_base_size: 1024 step: 128 min_size: 384 do_upscale: true caption_processors: - type: shuffle split_separator: "," optimizer: name: "schedulefree.RAdamScheduleFree" # name: "bitsandbytes.optim.AdamW8bit" args: lr: 0.001 scheduler: # name: "torch.optim.lr_scheduler.ConstantLR" # args: {} tracker: project_name: "auraflow-pvc-1" loggers: - wandb saving: strategy: per_epochs: 0.25 per_steps: null save_last: true callbacks: - type: "hf_hub" # or "hf_hub" to push to hub name: "aura-pvc-2-" save_dir: "./output/pvc-2" hub_id: "p1atdev/afv03-lora" dir_in_repo: "pvc-2" seed: 42 num_train_epochs: 10 trainer: # debug_mode: "1step" gradient_checkpointing: true torch_compile: true torch_compile_args: mode: max-autotune fullgraph: true fp32_matmul_precision: "medium"