afv03-lora / pvc-2 /config.yml
p1atdev's picture
Upload config.yml
d75e886 verified
raw
history blame
1.34 kB
model:
checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors"
pretrained_model_name_or_path: fal/AuraFlow-v0.3
dtype: bfloat16
denoiser:
use_flash_attn: true
peft:
type: lora
rank: 16
alpha: 1.0
dropout: 0.0
dtype: bfloat16
include_keys: [".attn.", ".mlp.", ".modC.", ".modC.", ".modX."]
exclude_keys: ["text_encoder", "vae", "t_embedder", "final_linear"]
dataset:
folder: "data/pvc"
num_repeats: 2
batch_size: 2
bucket_base_size: 1024
step: 128
min_size: 384
do_upscale: true
caption_processors:
- type: shuffle
split_separator: ","
optimizer:
name: "schedulefree.RAdamScheduleFree"
# name: "bitsandbytes.optim.AdamW8bit"
args:
lr: 0.001
scheduler:
# name: "torch.optim.lr_scheduler.ConstantLR"
# args: {}
tracker:
project_name: "auraflow-pvc-1"
loggers:
- wandb
saving:
strategy:
per_epochs: 0.25
per_steps: null
save_last: true
callbacks:
- type: "hf_hub" # or "hf_hub" to push to hub
name: "aura-pvc-2-"
save_dir: "./output/pvc-2"
hub_id: "p1atdev/afv03-lora"
dir_in_repo: "pvc-2"
seed: 42
num_train_epochs: 10
trainer:
# debug_mode: "1step"
gradient_checkpointing: true
torch_compile: true
torch_compile_args:
mode: max-autotune
fullgraph: true
fp32_matmul_precision: "medium"