patrickvonplaten commited on
Commit
b4034ce
1 Parent(s): 4e2e12c
scheduler/scheduler_config.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "_class_name": "VQDiffusionScheduler",
3
  "_diffusers_version": "0.7.0.dev0",
4
- "a_cumulative_end": 9e-06,
5
- "a_cumulative_start": 0.99999,
6
- "c_cumulative_end": 0.99999,
7
- "c_cumulative_start": 9e-06,
8
- "num_embed": 4097,
9
- "num_train_timesteps": 100
10
  }
 
1
  {
2
  "_class_name": "VQDiffusionScheduler",
3
  "_diffusers_version": "0.7.0.dev0",
4
+ "alpha_cum_end": 9e-06,
5
+ "alpha_cum_start": 0.99999,
6
+ "gamma_cum_end": 0.99999,
7
+ "gamma_cum_start": 9e-06,
8
+ "num_train_timesteps": 100,
9
+ "num_vec_classes": 4097
10
  }
transformer/config.json CHANGED
@@ -1,28 +1,16 @@
1
  {
2
  "_class_name": "Transformer2DModel",
3
  "_diffusers_version": "0.7.0.dev0",
 
4
  "attention_bias": true,
5
- "context_dim": 512,
6
- "d_head": 88,
7
- "depth": 36,
8
- "discrete": true,
9
  "dropout": 0.0,
10
- "ff_layers": [
11
- "Linear",
12
- "ApproximateGELU",
13
- "Linear",
14
- "Dropout"
15
- ],
16
- "height": 32,
17
  "in_channels": null,
18
- "n_heads": 16,
19
- "norm_layers": [
20
- "AdaLayerNorm",
21
- "AdaLayerNorm",
22
- "LayerNorm"
23
- ],
24
- "num_embed": 4097,
25
  "num_embeds_ada_norm": 100,
26
- "num_groups": 32,
27
- "width": 32
 
28
  }
 
1
  {
2
  "_class_name": "Transformer2DModel",
3
  "_diffusers_version": "0.7.0.dev0",
4
+ "activation_fn": "geglu-approximate",
5
  "attention_bias": true,
6
+ "attention_head_dim": 88,
7
+ "cross_attention_dim": 512,
 
 
8
  "dropout": 0.0,
 
 
 
 
 
 
 
9
  "in_channels": null,
10
+ "norm_num_groups": 32,
11
+ "num_attention_heads": 16,
 
 
 
 
 
12
  "num_embeds_ada_norm": 100,
13
+ "num_layers": 36,
14
+ "num_vector_embeds": 4097,
15
+ "sample_size": 32
16
  }
transformer/diffusion_pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b4f3633099b4dcf09ff0187cee3eb6ff3f918fd4f7d1a3dede87ef831b2f277
3
- size 5437498090
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:971c77397048f1473f321bf762c462b181b541b58616721f06d96ec04c00e313
3
+ size 5437498666
vqvae/config.json CHANGED
@@ -14,7 +14,6 @@
14
  "DownEncoderBlock2D",
15
  "AttnDownEncoderBlock2D"
16
  ],
17
- "e_dim": 128,
18
  "in_channels": 3,
19
  "latent_channels": 256,
20
  "layers_per_block": 2,
@@ -27,5 +26,6 @@
27
  "UpDecoderBlock2D",
28
  "UpDecoderBlock2D",
29
  "UpDecoderBlock2D"
30
- ]
 
31
  }
 
14
  "DownEncoderBlock2D",
15
  "AttnDownEncoderBlock2D"
16
  ],
 
17
  "in_channels": 3,
18
  "latent_channels": 256,
19
  "layers_per_block": 2,
 
26
  "UpDecoderBlock2D",
27
  "UpDecoderBlock2D",
28
  "UpDecoderBlock2D"
29
+ ],
30
+ "vq_embed_dim": 128
31
  }