Phospy commited on
Commit
8228780
·
1 Parent(s): e4a5461

feat: upload chise3 lora model

Browse files
pro/chise3/chise3-000010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa7796861fd7c4180e37c9be0253116abd0063e2d18c2c4cde2baa13d09d22b8
3
+ size 75619392
pro/chise3/chise3-000020.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cf767e7e0588052863c5cde6ad9f8ec2de9a32b3089ff8bd98115f7bfce6ef8
3
+ size 75619393
pro/chise3/chise3-000030.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0192e6c24b757e5a0ad082b25747ddb06a988962f768681eefd22a5352dbc2f2
3
+ size 75619393
pro/chise3/chise3-000040.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f3be96961749095c14117fe867ae3e416f402bda2020ad53a54bb8b3c3df57d
3
+ size 75619393
pro/chise3/chise3.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6157c42bf9a4a46074ec9485915bd8e5dbe894fffc62bc28bd4031df4356af58
3
+ size 75619393
pro/chise3/config_file.toml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [model_arguments]
2
+ v2 = false
3
+ v_parameterization = false
4
+ pretrained_model_name_or_path = "/content/pretrained_model/Animefull-final-pruned.ckpt"
5
+
6
+ [additional_network_arguments]
7
+ no_metadata = false
8
+ unet_lr = 1.0
9
+ text_encoder_lr = 1.0
10
+ network_module = "networks.lora"
11
+ network_dim = 64
12
+ network_alpha = 1
13
+ network_train_unet_only = false
14
+ network_train_text_encoder_only = false
15
+
16
+ [optimizer_arguments]
17
+ optimizer_type = "Prodigy"
18
+ learning_rate = 1.0
19
+ max_grad_norm = 1.0
20
+ optimizer_args = [ "decouple=True", "weight_decay=0.01", "d_coef=2", "use_bias_correction=True", "safeguard_warmup=True", "betas=0.9,0.99",]
21
+ lr_scheduler = "constant_with_warmup"
22
+ lr_warmup_steps = 100
23
+
24
+ [dataset_arguments]
25
+ cache_latents = true
26
+ debug_dataset = false
27
+ vae_batch_size = 4
28
+
29
+ [training_arguments]
30
+ output_dir = "/content/LoRA/output"
31
+ output_name = "chise3"
32
+ save_precision = "fp16"
33
+ save_every_n_epochs = 10
34
+ train_batch_size = 6
35
+ max_token_length = 225
36
+ mem_eff_attn = false
37
+ xformers = true
38
+ max_train_epochs = 50
39
+ max_data_loader_n_workers = 8
40
+ persistent_data_loader_workers = true
41
+ seed = 31337
42
+ gradient_checkpointing = false
43
+ gradient_accumulation_steps = 1
44
+ mixed_precision = "fp16"
45
+ clip_skip = 2
46
+ logging_dir = "/content/LoRA/logs"
47
+ log_prefix = "chise3"
48
+ lowram = true
49
+
50
+ [sample_prompt_arguments]
51
+ sample_every_n_epochs = 5
52
+ sample_sampler = "ddim"
53
+
54
+ [dreambooth_arguments]
55
+ prior_loss_weight = 1.0
56
+
57
+ [saving_arguments]
58
+ save_model_as = "safetensors"
pro/chise3/sample/chise3_20230822112423_e000005_00.png ADDED
pro/chise3/sample/chise3_20230822113805_e000010_00.png ADDED
pro/chise3/sample/chise3_20230822115146_e000015_00.png ADDED
pro/chise3/sample/chise3_20230822120527_e000020_00.png ADDED
pro/chise3/sample/chise3_20230822121910_e000025_00.png ADDED
pro/chise3/sample/chise3_20230822123256_e000030_00.png ADDED
pro/chise3/sample/chise3_20230822124639_e000035_00.png ADDED
pro/chise3/sample/chise3_20230822130024_e000040_00.png ADDED
pro/chise3/sample/chise3_20230822131406_e000045_00.png ADDED
pro/chise3/sample/chise3_20230822132750_e000050_00.png ADDED