panopstor commited on
Commit
875f115
1 Parent(s): 2f0204a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -12,4 +12,6 @@ Approximate time to train is 60 hours on an RTX 6000 Ada 48GB, initially trained
12
  Unet weights tuned with bitsandbytes AdamW8bit with 1e-6 learning rate, with constant learning rate for the first 30 epochs and cosine for the final 10.
13
  Text encoder tuned via backpropogation through the Unet using the same AdamW8bit optimizer with a learning rate of 2e-7 cosine for each session and weight decay 0.040 to account for lower normal.
14
 
15
- EveryDream2 trainer implements aspect ratio batch fitting so cropping artifacts are greatly reduced. Higher resolution outputs are consistent compared to the original SD1.5 checkpoint which tends to duplicate subject matter beyond the trained 512x512 resolution.
 
 
 
12
  Unet weights tuned with bitsandbytes AdamW8bit with 1e-6 learning rate, with constant learning rate for the first 30 epochs and cosine for the final 10.
13
  Text encoder tuned via backpropogation through the Unet using the same AdamW8bit optimizer with a learning rate of 2e-7 cosine for each session and weight decay 0.040 to account for lower normal.
14
 
15
+ EveryDream2 trainer implements aspect ratio batch fitting so cropping artifacts are greatly reduced. Higher resolution outputs are consistent compared to the original SD1.5 checkpoint which tends to duplicate subject matter beyond the trained 512x512 resolution.
16
+
17
+ Optimizer states are provided in .pt form for each the text encoder and Unet, which will can be loaded along with the Diffusers weights for resumption in the EveryDream2 trainer software.