potatp commited on
Commit
9158e2a
·
verified ·
1 Parent(s): 914b4d2

End of training

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +39 -0
  2. checkpoint-1000/model.safetensors +3 -0
  3. checkpoint-1000/optimizer.bin +3 -0
  4. checkpoint-1000/random_states_0.pkl +3 -0
  5. checkpoint-1000/scheduler.bin +3 -0
  6. checkpoint-1500/model.safetensors +3 -0
  7. checkpoint-1500/optimizer.bin +3 -0
  8. checkpoint-1500/random_states_0.pkl +3 -0
  9. checkpoint-1500/scheduler.bin +3 -0
  10. checkpoint-2000/model.safetensors +3 -0
  11. checkpoint-2000/optimizer.bin +3 -0
  12. checkpoint-2000/random_states_0.pkl +3 -0
  13. checkpoint-2000/scheduler.bin +3 -0
  14. checkpoint-2500/model.safetensors +3 -0
  15. checkpoint-2500/optimizer.bin +3 -0
  16. checkpoint-2500/random_states_0.pkl +3 -0
  17. checkpoint-2500/scheduler.bin +3 -0
  18. checkpoint-3000/model.safetensors +3 -0
  19. checkpoint-3000/optimizer.bin +3 -0
  20. checkpoint-3000/random_states_0.pkl +3 -0
  21. checkpoint-3000/scheduler.bin +3 -0
  22. checkpoint-500/model.safetensors +3 -0
  23. checkpoint-500/optimizer.bin +3 -0
  24. checkpoint-500/random_states_0.pkl +3 -0
  25. checkpoint-500/scheduler.bin +3 -0
  26. feature_extractor/preprocessor_config.json +27 -0
  27. learned_embeds-steps-1000.safetensors +3 -0
  28. learned_embeds-steps-1500.safetensors +3 -0
  29. learned_embeds-steps-2000.safetensors +3 -0
  30. learned_embeds-steps-2500.safetensors +3 -0
  31. learned_embeds-steps-3000.safetensors +3 -0
  32. learned_embeds-steps-500.safetensors +3 -0
  33. learned_embeds.safetensors +3 -0
  34. logs/textual_inversion/1731236974.8500457/events.out.tfevents.1731236974.WIN-9SR36LMNDB2.518335.1 +3 -0
  35. logs/textual_inversion/1731236974.8613741/hparams.yml +48 -0
  36. logs/textual_inversion/events.out.tfevents.1731236974.WIN-9SR36LMNDB2.518335.0 +3 -0
  37. model_index.json +38 -0
  38. safety_checker/config.json +29 -0
  39. safety_checker/model.safetensors +3 -0
  40. scheduler/scheduler_config.json +15 -0
  41. text_encoder/config.json +25 -0
  42. text_encoder/model.safetensors +3 -0
  43. tokenizer/added_tokens.json +22 -0
  44. tokenizer/merges.txt +0 -0
  45. tokenizer/special_tokens_map.json +24 -0
  46. tokenizer/tokenizer_config.json +190 -0
  47. tokenizer/vocab.json +0 -0
  48. unet/config.json +68 -0
  49. unet/diffusion_pytorch_model.safetensors +3 -0
  50. vae/config.json +38 -0
README.md ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: stable-diffusion-v1-5/stable-diffusion-v1-5
3
+ library_name: diffusers
4
+ license: creativeml-openrail-m
5
+ inference: true
6
+ tags:
7
+ - stable-diffusion
8
+ - stable-diffusion-diffusers
9
+ - text-to-image
10
+ - diffusers
11
+ - textual_inversion
12
+ - diffusers-training
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the training script had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+
19
+ # Textual inversion text2image fine-tuning - potatp/textual_inversion_screw3
20
+ These are textual inversion adaption weights for stable-diffusion-v1-5/stable-diffusion-v1-5. You can find some example images in the following.
21
+
22
+
23
+
24
+
25
+ ## Intended uses & limitations
26
+
27
+ #### How to use
28
+
29
+ ```python
30
+ # TODO: add an example code snippet for running this diffusion pipeline
31
+ ```
32
+
33
+ #### Limitations and bias
34
+
35
+ [TODO: provide examples of latent issues and potential remediations]
36
+
37
+ ## Training details
38
+
39
+ [TODO: describe the data used to train the model]
checkpoint-1000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ca7ba67bfbe239a99b209dc47cea54b27f7363fded15741938f4a7155410344
3
+ size 492326608
checkpoint-1000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d41a40ba5f746b25e234807f25f9934859bd7eb3f99d71d7f02596ac770a58d3
3
+ size 303687650
checkpoint-1000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:788e4f495eb0553b41cf7872f5b5ed685f01b9a1b60a03dc5fc4ebabd71f5b6e
3
+ size 14408
checkpoint-1000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f0d829441992ff2ab1f7502a3a94afc011706461052e42e18e5b963678527c4
3
+ size 1000
checkpoint-1500/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:131c5d9b3fcbe0ebb2f67fff92a1225c2afa00b1c2ffd0ad4deec15a9672d93b
3
+ size 492326608
checkpoint-1500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b080a0c68ff8aa71e425e4ea9977485c26143fc87d19b0643f9b683db39ad4bd
3
+ size 303687650
checkpoint-1500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:397990f80dd12806fe29ef567a794aee95982197cc1badf75c30149bccb0d9d1
3
+ size 14408
checkpoint-1500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0746407d5e6c8d929c4e940db747c88272550cab75cfb856468297d71d7802ca
3
+ size 1000
checkpoint-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe14b412dd21780493f5441aeb2d7a9f7c1f10aae39bb8df789bf73a6501b66a
3
+ size 492326608
checkpoint-2000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:660d26bd0c3d37571821e28cd2f476ed84a52d2a629feb69fad1da4803c0cfa4
3
+ size 303687650
checkpoint-2000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea5d83abd55950232728265e2ccdc86c44bdc3fea341815facbe230babcfd37b
3
+ size 14344
checkpoint-2000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5484c6ff4bca55d86f0d54575371a1f3b26246b4ff863161800c9be8539d8c77
3
+ size 1000
checkpoint-2500/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d022a4dd8cc5c77806815ab507363ac25f1b63e0b305b39017bfeed9f376772
3
+ size 492326608
checkpoint-2500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11923a422da3c45a5d7ff1ae45f445fd043a26d8b6f434ba835b958a081bac8a
3
+ size 303687650
checkpoint-2500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a09349b3d872c6dc5b7f8924824340d5cb8b863092cfdf6ce8c0bb7973a0163
3
+ size 14344
checkpoint-2500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bb1a67b9696a2697693774a6f09d8186c78e37cff7896931a3632a6822f70b9
3
+ size 1000
checkpoint-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0b726c1796a0313547f562efb49a67e05ead4ebca5cf45b86304264cf279309
3
+ size 492326608
checkpoint-3000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e79ad957e30314c6bfaf99f19b3b1477e2ba7e35a48a53b89950a5a00081768
3
+ size 303687650
checkpoint-3000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aca6b7a55b330f53291c1f602c3aee92c8a0964f34e6a3e631766378df93608a
3
+ size 14344
checkpoint-3000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ed1284c2bacec58dd33ce8bc7070e7691d6d084b54330b85bacc37784d59cee
3
+ size 1000
checkpoint-500/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ff18da381da1ba9649b2b382c0853fb30f9098d15811e5ec3cc13e2052bb6ec
3
+ size 492326608
checkpoint-500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:848f173d661cf152088b25c75f730678e08acde2760acee7d49416daea124eb4
3
+ size 303687650
checkpoint-500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:915c30ef7a598f6c937910e8ea95728dd90758aeb4eaeab11493b8fb96b526d9
3
+ size 14280
checkpoint-500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cad514bb36617120f19b7cde3b36a22bc8becf6cc6cfb2166811645c215cccb
3
+ size 1000
feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "image_mean": [
12
+ 0.48145466,
13
+ 0.4578275,
14
+ 0.40821073
15
+ ],
16
+ "image_processor_type": "CLIPImageProcessor",
17
+ "image_std": [
18
+ 0.26862954,
19
+ 0.26130258,
20
+ 0.27577711
21
+ ],
22
+ "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "shortest_edge": 224
26
+ }
27
+ }
learned_embeds-steps-1000.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e804304061322b521ead0c4b61f70afa3173f3d1cd3809cb224ca2f2820d1d21
3
+ size 61560
learned_embeds-steps-1500.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bdffd8b3555b8ad3cd1d5d202d5b0ad9d4dee9d7af45a821aa84029e4140a51
3
+ size 61560
learned_embeds-steps-2000.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e90791e91bafaa3fcdf3c4db3eb0d33fade199294bf753fdac6394db0da8f6d
3
+ size 61560
learned_embeds-steps-2500.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e71f09cbc62add3672101cfff3d9ac943bb7169f0eb56328dcb535b1ff17ed35
3
+ size 61560
learned_embeds-steps-3000.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e524562ddec514498ab11f2fb9bd69e5cbcfbe787aa0a1299f27ac5962d45544
3
+ size 61560
learned_embeds-steps-500.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11e52080aa9a111c471b7d08df9e23da56fdbcbc5d6bae179ff7600f85af2765
3
+ size 61560
learned_embeds.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e524562ddec514498ab11f2fb9bd69e5cbcfbe787aa0a1299f27ac5962d45544
3
+ size 61560
logs/textual_inversion/1731236974.8500457/events.out.tfevents.1731236974.WIN-9SR36LMNDB2.518335.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7788c8f73ebbd02db6c835a6fec841623d21ff508c170d12e0cae9786d6b72d7
3
+ size 2348
logs/textual_inversion/1731236974.8613741/hparams.yml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ center_crop: false
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ enable_xformers_memory_efficient_attention: false
11
+ gradient_accumulation_steps: 4
12
+ gradient_checkpointing: false
13
+ hub_model_id: null
14
+ hub_token: null
15
+ initializer_token: metal
16
+ learnable_property: object
17
+ learning_rate: 0.002
18
+ local_rank: -1
19
+ logging_dir: logs
20
+ lr_num_cycles: 1
21
+ lr_scheduler: constant
22
+ lr_warmup_steps: 0
23
+ max_train_steps: 3000
24
+ mixed_precision: 'no'
25
+ no_safe_serialization: false
26
+ num_train_epochs: 24
27
+ num_validation_images: 4
28
+ num_vectors: 20
29
+ output_dir: textual_inversion_screw3
30
+ placeholder_token: <screw-metal>
31
+ pretrained_model_name_or_path: stable-diffusion-v1-5/stable-diffusion-v1-5
32
+ push_to_hub: true
33
+ repeats: 100
34
+ report_to: tensorboard
35
+ resolution: 512
36
+ resume_from_checkpoint: null
37
+ revision: null
38
+ save_as_full_pipeline: false
39
+ save_steps: 500
40
+ scale_lr: true
41
+ seed: null
42
+ tokenizer_name: null
43
+ train_batch_size: 1
44
+ train_data_dir: data_dir
45
+ validation_epochs: null
46
+ validation_prompt: null
47
+ validation_steps: 100
48
+ variant: null
logs/textual_inversion/events.out.tfevents.1731236974.WIN-9SR36LMNDB2.518335.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f89316135248d4316108b1de5bfe5661bdff52b3d1ede2ec852c19c212ba3a59
3
+ size 1007054
model_index.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.32.0.dev0",
4
+ "_name_or_path": "stable-diffusion-v1-5/stable-diffusion-v1-5",
5
+ "feature_extractor": [
6
+ "transformers",
7
+ "CLIPImageProcessor"
8
+ ],
9
+ "image_encoder": [
10
+ null,
11
+ null
12
+ ],
13
+ "requires_safety_checker": true,
14
+ "safety_checker": [
15
+ "stable_diffusion",
16
+ "StableDiffusionSafetyChecker"
17
+ ],
18
+ "scheduler": [
19
+ "diffusers",
20
+ "PNDMScheduler"
21
+ ],
22
+ "text_encoder": [
23
+ "transformers",
24
+ "CLIPTextModel"
25
+ ],
26
+ "tokenizer": [
27
+ "transformers",
28
+ "CLIPTokenizer"
29
+ ],
30
+ "unet": [
31
+ "diffusers",
32
+ "UNet2DConditionModel"
33
+ ],
34
+ "vae": [
35
+ "diffusers",
36
+ "AutoencoderKL"
37
+ ]
38
+ }
safety_checker/config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/home/potato/.cache/huggingface/hub/models--stable-diffusion-v1-5--stable-diffusion-v1-5/snapshots/f03de327dd89b501a01da37fc5240cf4fdba85a1/safety_checker",
3
+ "architectures": [
4
+ "StableDiffusionSafetyChecker"
5
+ ],
6
+ "initializer_factor": 1.0,
7
+ "logit_scale_init_value": 2.6592,
8
+ "model_type": "clip",
9
+ "projection_dim": 768,
10
+ "text_config": {
11
+ "dropout": 0.0,
12
+ "hidden_size": 768,
13
+ "intermediate_size": 3072,
14
+ "model_type": "clip_text_model",
15
+ "num_attention_heads": 12
16
+ },
17
+ "torch_dtype": "float32",
18
+ "transformers_version": "4.46.2",
19
+ "vision_config": {
20
+ "_attn_implementation_autoset": true,
21
+ "dropout": 0.0,
22
+ "hidden_size": 1024,
23
+ "intermediate_size": 4096,
24
+ "model_type": "clip_vision_model",
25
+ "num_attention_heads": 16,
26
+ "num_hidden_layers": 24,
27
+ "patch_size": 14
28
+ }
29
+ }
safety_checker/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb351a5ded815c3ff744968ad9c6b218d071b9d313d04f35e813b84b4c0ffde8
3
+ size 1215979664
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.32.0.dev0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
+ "set_alpha_to_one": false,
11
+ "skip_prk_steps": true,
12
+ "steps_offset": 1,
13
+ "timestep_spacing": "leading",
14
+ "trained_betas": null
15
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "stable-diffusion-v1-5/stable-diffusion-v1-5",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 768,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.46.2",
24
+ "vocab_size": 49428
25
+ }
text_encoder/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0b726c1796a0313547f562efb49a67e05ead4ebca5cf45b86304264cf279309
3
+ size 492326608
tokenizer/added_tokens.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<screw-metal>": 49408,
3
+ "<screw-metal>_1": 49409,
4
+ "<screw-metal>_10": 49418,
5
+ "<screw-metal>_11": 49419,
6
+ "<screw-metal>_12": 49420,
7
+ "<screw-metal>_13": 49421,
8
+ "<screw-metal>_14": 49422,
9
+ "<screw-metal>_15": 49423,
10
+ "<screw-metal>_16": 49424,
11
+ "<screw-metal>_17": 49425,
12
+ "<screw-metal>_18": 49426,
13
+ "<screw-metal>_19": 49427,
14
+ "<screw-metal>_2": 49410,
15
+ "<screw-metal>_3": 49411,
16
+ "<screw-metal>_4": 49412,
17
+ "<screw-metal>_5": 49413,
18
+ "<screw-metal>_6": 49414,
19
+ "<screw-metal>_7": 49415,
20
+ "<screw-metal>_8": 49416,
21
+ "<screw-metal>_9": 49417
22
+ }
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "49406": {
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "49407": {
13
+ "content": "<|endoftext|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "49408": {
21
+ "content": "<screw-metal>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": false
27
+ },
28
+ "49409": {
29
+ "content": "<screw-metal>_1",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": false
35
+ },
36
+ "49410": {
37
+ "content": "<screw-metal>_2",
38
+ "lstrip": false,
39
+ "normalized": true,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": false
43
+ },
44
+ "49411": {
45
+ "content": "<screw-metal>_3",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": false
51
+ },
52
+ "49412": {
53
+ "content": "<screw-metal>_4",
54
+ "lstrip": false,
55
+ "normalized": true,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": false
59
+ },
60
+ "49413": {
61
+ "content": "<screw-metal>_5",
62
+ "lstrip": false,
63
+ "normalized": true,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": false
67
+ },
68
+ "49414": {
69
+ "content": "<screw-metal>_6",
70
+ "lstrip": false,
71
+ "normalized": true,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": false
75
+ },
76
+ "49415": {
77
+ "content": "<screw-metal>_7",
78
+ "lstrip": false,
79
+ "normalized": true,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": false
83
+ },
84
+ "49416": {
85
+ "content": "<screw-metal>_8",
86
+ "lstrip": false,
87
+ "normalized": true,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": false
91
+ },
92
+ "49417": {
93
+ "content": "<screw-metal>_9",
94
+ "lstrip": false,
95
+ "normalized": true,
96
+ "rstrip": false,
97
+ "single_word": false,
98
+ "special": false
99
+ },
100
+ "49418": {
101
+ "content": "<screw-metal>_10",
102
+ "lstrip": false,
103
+ "normalized": true,
104
+ "rstrip": false,
105
+ "single_word": false,
106
+ "special": false
107
+ },
108
+ "49419": {
109
+ "content": "<screw-metal>_11",
110
+ "lstrip": false,
111
+ "normalized": true,
112
+ "rstrip": false,
113
+ "single_word": false,
114
+ "special": false
115
+ },
116
+ "49420": {
117
+ "content": "<screw-metal>_12",
118
+ "lstrip": false,
119
+ "normalized": true,
120
+ "rstrip": false,
121
+ "single_word": false,
122
+ "special": false
123
+ },
124
+ "49421": {
125
+ "content": "<screw-metal>_13",
126
+ "lstrip": false,
127
+ "normalized": true,
128
+ "rstrip": false,
129
+ "single_word": false,
130
+ "special": false
131
+ },
132
+ "49422": {
133
+ "content": "<screw-metal>_14",
134
+ "lstrip": false,
135
+ "normalized": true,
136
+ "rstrip": false,
137
+ "single_word": false,
138
+ "special": false
139
+ },
140
+ "49423": {
141
+ "content": "<screw-metal>_15",
142
+ "lstrip": false,
143
+ "normalized": true,
144
+ "rstrip": false,
145
+ "single_word": false,
146
+ "special": false
147
+ },
148
+ "49424": {
149
+ "content": "<screw-metal>_16",
150
+ "lstrip": false,
151
+ "normalized": true,
152
+ "rstrip": false,
153
+ "single_word": false,
154
+ "special": false
155
+ },
156
+ "49425": {
157
+ "content": "<screw-metal>_17",
158
+ "lstrip": false,
159
+ "normalized": true,
160
+ "rstrip": false,
161
+ "single_word": false,
162
+ "special": false
163
+ },
164
+ "49426": {
165
+ "content": "<screw-metal>_18",
166
+ "lstrip": false,
167
+ "normalized": true,
168
+ "rstrip": false,
169
+ "single_word": false,
170
+ "special": false
171
+ },
172
+ "49427": {
173
+ "content": "<screw-metal>_19",
174
+ "lstrip": false,
175
+ "normalized": true,
176
+ "rstrip": false,
177
+ "single_word": false,
178
+ "special": false
179
+ }
180
+ },
181
+ "bos_token": "<|startoftext|>",
182
+ "clean_up_tokenization_spaces": false,
183
+ "do_lower_case": true,
184
+ "eos_token": "<|endoftext|>",
185
+ "errors": "replace",
186
+ "model_max_length": 77,
187
+ "pad_token": "<|endoftext|>",
188
+ "tokenizer_class": "CLIPTokenizer",
189
+ "unk_token": "<|endoftext|>"
190
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.32.0.dev0",
4
+ "_name_or_path": "stable-diffusion-v1-5/stable-diffusion-v1-5",
5
+ "act_fn": "silu",
6
+ "addition_embed_type": null,
7
+ "addition_embed_type_num_heads": 64,
8
+ "addition_time_embed_dim": null,
9
+ "attention_head_dim": 8,
10
+ "attention_type": "default",
11
+ "block_out_channels": [
12
+ 320,
13
+ 640,
14
+ 1280,
15
+ 1280
16
+ ],
17
+ "center_input_sample": false,
18
+ "class_embed_type": null,
19
+ "class_embeddings_concat": false,
20
+ "conv_in_kernel": 3,
21
+ "conv_out_kernel": 3,
22
+ "cross_attention_dim": 768,
23
+ "cross_attention_norm": null,
24
+ "down_block_types": [
25
+ "CrossAttnDownBlock2D",
26
+ "CrossAttnDownBlock2D",
27
+ "CrossAttnDownBlock2D",
28
+ "DownBlock2D"
29
+ ],
30
+ "downsample_padding": 1,
31
+ "dropout": 0.0,
32
+ "dual_cross_attention": false,
33
+ "encoder_hid_dim": null,
34
+ "encoder_hid_dim_type": null,
35
+ "flip_sin_to_cos": true,
36
+ "freq_shift": 0,
37
+ "in_channels": 4,
38
+ "layers_per_block": 2,
39
+ "mid_block_only_cross_attention": null,
40
+ "mid_block_scale_factor": 1,
41
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
42
+ "norm_eps": 1e-05,
43
+ "norm_num_groups": 32,
44
+ "num_attention_heads": null,
45
+ "num_class_embeds": null,
46
+ "only_cross_attention": false,
47
+ "out_channels": 4,
48
+ "projection_class_embeddings_input_dim": null,
49
+ "resnet_out_scale_factor": 1.0,
50
+ "resnet_skip_time_act": false,
51
+ "resnet_time_scale_shift": "default",
52
+ "reverse_transformer_layers_per_block": null,
53
+ "sample_size": 64,
54
+ "time_cond_proj_dim": null,
55
+ "time_embedding_act_fn": null,
56
+ "time_embedding_dim": null,
57
+ "time_embedding_type": "positional",
58
+ "timestep_post_act": null,
59
+ "transformer_layers_per_block": 1,
60
+ "up_block_types": [
61
+ "UpBlock2D",
62
+ "CrossAttnUpBlock2D",
63
+ "CrossAttnUpBlock2D",
64
+ "CrossAttnUpBlock2D"
65
+ ],
66
+ "upcast_attention": false,
67
+ "use_linear_projection": false
68
+ }
unet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d27cd69d4a0aa32105087a619f32a51bc087e133be93fe23da92f3c0bcc07d79
3
+ size 3438167536
vae/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.32.0.dev0",
4
+ "_name_or_path": "stable-diffusion-v1-5/stable-diffusion-v1-5",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "force_upcast": true,
19
+ "in_channels": 3,
20
+ "latent_channels": 4,
21
+ "latents_mean": null,
22
+ "latents_std": null,
23
+ "layers_per_block": 2,
24
+ "mid_block_add_attention": true,
25
+ "norm_num_groups": 32,
26
+ "out_channels": 3,
27
+ "sample_size": 512,
28
+ "scaling_factor": 0.18215,
29
+ "shift_factor": null,
30
+ "up_block_types": [
31
+ "UpDecoderBlock2D",
32
+ "UpDecoderBlock2D",
33
+ "UpDecoderBlock2D",
34
+ "UpDecoderBlock2D"
35
+ ],
36
+ "use_post_quant_conv": true,
37
+ "use_quant_conv": true
38
+ }