Upload StableDiffusion3Pipeline
Browse files- text_encoder/config.json +1 -0
- text_encoder_2/config.json +1 -0
- text_encoder_3/config.json +1 -0
- transformer/config.json +1 -0
- transformer/diffusion_pytorch_model.safetensors +2 -2
- vae/config.json +1 -0
text_encoder/config.json
CHANGED
@@ -1,4 +1,5 @@
|
|
1 |
{
|
|
|
2 |
"architectures": [
|
3 |
"CLIPTextModelWithProjection"
|
4 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/host/ssd2/yamaguchi-r/.cache/huggingface/hub/models--stabilityai--stable-diffusion-3.5-medium/snapshots/b940f670f0eda2d07fbb75229e779da1ad11eb80/text_encoder",
|
3 |
"architectures": [
|
4 |
"CLIPTextModelWithProjection"
|
5 |
],
|
text_encoder_2/config.json
CHANGED
@@ -1,4 +1,5 @@
|
|
1 |
{
|
|
|
2 |
"architectures": [
|
3 |
"CLIPTextModelWithProjection"
|
4 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/host/ssd2/yamaguchi-r/.cache/huggingface/hub/models--stabilityai--stable-diffusion-3.5-medium/snapshots/b940f670f0eda2d07fbb75229e779da1ad11eb80/text_encoder_2",
|
3 |
"architectures": [
|
4 |
"CLIPTextModelWithProjection"
|
5 |
],
|
text_encoder_3/config.json
CHANGED
@@ -1,4 +1,5 @@
|
|
1 |
{
|
|
|
2 |
"architectures": [
|
3 |
"T5EncoderModel"
|
4 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/host/ssd2/yamaguchi-r/.cache/huggingface/hub/models--stabilityai--stable-diffusion-3.5-medium/snapshots/b940f670f0eda2d07fbb75229e779da1ad11eb80/text_encoder_3",
|
3 |
"architectures": [
|
4 |
"T5EncoderModel"
|
5 |
],
|
transformer/config.json
CHANGED
@@ -1,6 +1,7 @@
|
|
1 |
{
|
2 |
"_class_name": "SD3Transformer2DModel",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
|
|
4 |
"attention_head_dim": 64,
|
5 |
"caption_projection_dim": 1536,
|
6 |
"dual_attention_layers": [
|
|
|
1 |
{
|
2 |
"_class_name": "SD3Transformer2DModel",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
4 |
+
"_name_or_path": "/host/ssd2/yamaguchi-r/.cache/huggingface/hub/models--stabilityai--stable-diffusion-3.5-medium/snapshots/b940f670f0eda2d07fbb75229e779da1ad11eb80/transformer",
|
5 |
"attention_head_dim": 64,
|
6 |
"caption_projection_dim": 1536,
|
7 |
"dual_attention_layers": [
|
transformer/diffusion_pytorch_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c09165eefd07231b790e986b395357dab72f325f00ecaa8c3aebf101dfdbe7fd
|
3 |
+
size 4939432760
|
vae/config.json
CHANGED
@@ -1,6 +1,7 @@
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
|
|
4 |
"act_fn": "silu",
|
5 |
"block_out_channels": [
|
6 |
128,
|
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
"_diffusers_version": "0.32.0.dev0",
|
4 |
+
"_name_or_path": "/host/ssd2/yamaguchi-r/.cache/huggingface/hub/models--stabilityai--stable-diffusion-3.5-medium/snapshots/b940f670f0eda2d07fbb75229e779da1ad11eb80/vae",
|
5 |
"act_fn": "silu",
|
6 |
"block_out_channels": [
|
7 |
128,
|