model: checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors" pretrained_model_name_or_path: fal/AuraFlow-v0.3 dtype: bfloat16 denoiser: use_flash_attn: true use_rope: True rope_theta: 10000 rope_dim_sizes: [32, 112, 112] peft: type: lora rank: 4 alpha: 1.0 dropout: 0.0 dtype: bfloat16 include_keys: [".attn."] exclude_keys: ["text_encoder", "vae", "t_embedder", "final_linear"] dataset: folder: "data/pexels-1k-random" num_repeats: 2 batch_size: 2 bucket_base_size: 1024 step: 128 min_size: 384 caption_processors: [] optimizer: name: "schedulefree.RAdamScheduleFree" # name: "bitsandbytes.optim.AdamW8bit" args: lr: 0.005 scheduler: # name: "torch.optim.lr_scheduler.ConstantLR" # args: {} tracker: project_name: "auraflow-rope-1" loggers: - wandb saving: strategy: per_epochs: 1 per_steps: null save_last: true callbacks: - type: "hf_hub" # or "hf_hub" to push to hub name: "rope-4" save_dir: "./output/rope-4" hub_id: "p1atdev/afv03-lora" dir_in_repo: "rope-4" seed: 42 num_train_epochs: 5 trainer: # debug_mode: "1step" gradient_checkpointing: true torch_compile: true torch_compile_args: mode: max-autotune fullgraph: true fp32_matmul_precision: "medium"