Jingya HF staff commited on
Commit
89d090e
1 Parent(s): 5272eb2

Synchronizing local compiler cache.

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +223 -0
  2. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/clip_text_model/hf-internal-testing/tiny-stable-diffusion-torch/0058934baf38ad450d00.json +1 -0
  3. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/clip_text_model/hf-internal-testing/tiny-stable-diffusion-torch/1cd8c85de423d1a4d17f.json +1 -0
  4. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/clip_text_model/hf-internal-testing/tiny-stable-diffusion-torch/6437b1e4638636e96aa9.json +1 -0
  5. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-latent-consistency/3c97e4967e6c83d09039.json +1 -0
  6. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-latent-consistency/c0f11eca47e6ca3b4ca9.json +1 -0
  7. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/196075233ad72b81121c.json +1 -0
  8. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/5c62704e5907ab724408.json +1 -0
  9. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/7560f6c90b8190fc2cc9.json +1 -0
  10. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/7b816cfd640f458703b9.json +1 -0
  11. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/f5c5adf7a000b140312a.json +1 -0
  12. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/0834c9a64499cfd8014d.json +1 -0
  13. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/1e7de280b7fd1deb3ad9.json +1 -0
  14. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/5acab5b0d8d88e6c98af.json +1 -0
  15. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/6fbed240d612716ec707.json +1 -0
  16. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/a948b04ba5bae8fff61f.json +1 -0
  17. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/cd78efc0548f20d31d32.json +1 -0
  18. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/dba159d1c78cbf56d958.json +1 -0
  19. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/f58a4f1711d1bbd1723e.json +1 -0
  20. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/timbrooks/instruct-pix2pix/20f8a909335cd43da695.json +1 -0
  21. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/t5/hf-internal-testing/tiny-random-t5/15d0d730286c73e4f68e.json +1 -0
  22. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/facebook/wav2vec2-large-960h-lv60-self/56b22768a28f12bef9fb.json +1 -0
  23. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/facebook/wav2vec2-large-960h-lv60-self/86728684030d9f82c4d7.json +1 -0
  24. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/1f1c4a44a3fddd985756.json +1 -0
  25. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/3a64123a157e84d10501.json +1 -0
  26. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/472cb15e97bf2881387b.json +1 -0
  27. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/70bb5c3cf8828a56422c.json +1 -0
  28. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/76c77da34788c0ea062e.json +1 -0
  29. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/92a3aa68c4cd1c5e34d3.json +1 -0
  30. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/9b9771ddbdc9ec31c7d9.json +1 -0
  31. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/b012ed40f5e7d6b4bf0c.json +1 -0
  32. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/ce86b7c5d629e3d3a32c.json +1 -0
  33. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/db25ba0198786d9e0d25.json +1 -0
  34. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/dbbb34cc758fbdfd9620.json +1 -0
  35. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/e17307a51fd6bfd07860.json +1 -0
  36. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/e58341214d08e448b289.json +1 -0
  37. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/ed50ef1b7d06e90af218.json +1 -0
  38. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/feature_extractor/preprocessor_config.json +44 -0
  39. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/model_index.json +38 -0
  40. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/scheduler/scheduler_config.json +15 -0
  41. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/text_encoder/config.json +48 -0
  42. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/text_encoder/model.neuron +0 -0
  43. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/tokenizer/merges.txt +647 -0
  44. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/tokenizer/special_tokens_map.json +24 -0
  45. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/tokenizer/tokenizer_config.json +30 -0
  46. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/tokenizer/vocab.json +1002 -0
  47. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_decoder/config.json +65 -0
  48. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_decoder/model.neuron +3 -0
  49. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_encoder/config.json +65 -0
  50. neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_encoder/model.neuron +3 -0
.gitattributes CHANGED
@@ -4850,3 +4850,226 @@ neuronxcc-2.15.128.0+56dc5a86/MODULE_e7bd7a2134c7496f5ce9+39f12043/model.neff fi
4850
  neuronxcc-2.15.128.0+56dc5a86/MODULE_f8c98d89b4c3276b429e+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
4851
  neuronxcc-2.15.128.0+56dc5a86/MODULE_02e53ffc6085b74dd1c4+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
4852
  neuronxcc-2.15.128.0+56dc5a86/MODULE_e820688eff57f3ac8212+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4850
  neuronxcc-2.15.128.0+56dc5a86/MODULE_f8c98d89b4c3276b429e+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
4851
  neuronxcc-2.15.128.0+56dc5a86/MODULE_02e53ffc6085b74dd1c4+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
4852
  neuronxcc-2.15.128.0+56dc5a86/MODULE_e820688eff57f3ac8212+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
4853
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4854
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4855
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_0834c9a64499cfd8014d/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4856
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_0834c9a64499cfd8014d/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4857
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_0834c9a64499cfd8014d/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4858
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_0834c9a64499cfd8014d/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4859
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_196075233ad72b81121c/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4860
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_196075233ad72b81121c/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4861
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_196075233ad72b81121c/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4862
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1cd8c85de423d1a4d17f/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4863
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1cd8c85de423d1a4d17f/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4864
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1d36776db14b48f3353c/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4865
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1d36776db14b48f3353c/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4866
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1d36776db14b48f3353c/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4867
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1e7de280b7fd1deb3ad9/controlnet_0/model.neuron filter=lfs diff=lfs merge=lfs -text
4868
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1e7de280b7fd1deb3ad9/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4869
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1e7de280b7fd1deb3ad9/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4870
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1e7de280b7fd1deb3ad9/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4871
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_1e7de280b7fd1deb3ad9/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4872
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_20f8a909335cd43da695/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4873
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_20f8a909335cd43da695/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4874
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_20f8a909335cd43da695/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4875
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_20f8a909335cd43da695/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4876
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_2a358aadcc0e9b693ef5/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4877
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_2a358aadcc0e9b693ef5/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4878
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_2a358aadcc0e9b693ef5/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4879
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_3c97e4967e6c83d09039/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4880
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_3c97e4967e6c83d09039/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4881
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_3c97e4967e6c83d09039/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4882
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_3c97e4967e6c83d09039/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4883
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_56b22768a28f12bef9fb/model.neuron filter=lfs diff=lfs merge=lfs -text
4884
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_5acab5b0d8d88e6c98af/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4885
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_5acab5b0d8d88e6c98af/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4886
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_5acab5b0d8d88e6c98af/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4887
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_5acab5b0d8d88e6c98af/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4888
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_5c62704e5907ab724408/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4889
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_5c62704e5907ab724408/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4890
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_5c62704e5907ab724408/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4891
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_6437b1e4638636e96aa9/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4892
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_6437b1e4638636e96aa9/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4893
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_6fbed240d612716ec707/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4894
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_6fbed240d612716ec707/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4895
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_6fbed240d612716ec707/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4896
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_7560f6c90b8190fc2cc9/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4897
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_7560f6c90b8190fc2cc9/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4898
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_7560f6c90b8190fc2cc9/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4899
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_7b816cfd640f458703b9/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4900
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_7b816cfd640f458703b9/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4901
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_7b816cfd640f458703b9/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4902
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_86728684030d9f82c4d7/model.neuron filter=lfs diff=lfs merge=lfs -text
4903
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_a948b04ba5bae8fff61f/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4904
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_a948b04ba5bae8fff61f/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4905
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_a948b04ba5bae8fff61f/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4906
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_c0f11eca47e6ca3b4ca9/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4907
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_c0f11eca47e6ca3b4ca9/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4908
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_c0f11eca47e6ca3b4ca9/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4909
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_cd78efc0548f20d31d32/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4910
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_cd78efc0548f20d31d32/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4911
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_cd78efc0548f20d31d32/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4912
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_dba159d1c78cbf56d958/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4913
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_dba159d1c78cbf56d958/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4914
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_dba159d1c78cbf56d958/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4915
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_ef343604a8a1ee970660/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4916
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_ef343604a8a1ee970660/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4917
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_ef343604a8a1ee970660/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4918
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_f58a4f1711d1bbd1723e/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4919
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_f58a4f1711d1bbd1723e/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4920
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_f58a4f1711d1bbd1723e/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4921
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_f5c5adf7a000b140312a/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4922
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_f5c5adf7a000b140312a/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
4923
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_f5c5adf7a000b140312a/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4924
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_f5c5adf7a000b140312a/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4925
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_f5c5adf7a000b140312a/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4926
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_013d2a2858a7e279b962/controlnet_0/model.neuron filter=lfs diff=lfs merge=lfs -text
4927
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_09efc8a8af0f17d5b6d1/controlnet_0/model.neuron filter=lfs diff=lfs merge=lfs -text
4928
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_09efc8a8af0f17d5b6d1/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4929
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_09efc8a8af0f17d5b6d1/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4930
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_09efc8a8af0f17d5b6d1/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4931
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_09efc8a8af0f17d5b6d1/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4932
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_1c436c19018021ff6d90/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4933
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_1c436c19018021ff6d90/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
4934
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_1c436c19018021ff6d90/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4935
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_1c436c19018021ff6d90/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4936
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_1c436c19018021ff6d90/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4937
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_1fd1878703459f9fd034/decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4938
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_278599e82b0054f5776e/model.neuron filter=lfs diff=lfs merge=lfs -text
4939
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_298b2af089b860cb083e/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4940
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_298b2af089b860cb083e/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4941
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_298b2af089b860cb083e/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4942
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_3c9caaf25478da247ba9/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4943
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_3c9caaf25478da247ba9/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4944
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_3c9caaf25478da247ba9/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4945
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_429edd8d3e94cff265a9/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4946
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_429edd8d3e94cff265a9/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4947
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_429edd8d3e94cff265a9/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4948
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_5584e5d99a9cca114714/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4949
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_5584e5d99a9cca114714/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4950
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_5584e5d99a9cca114714/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4951
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_633a5bb695bb8f04bc7e/controlnet_0/model.neuron filter=lfs diff=lfs merge=lfs -text
4952
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_633a5bb695bb8f04bc7e/controlnet_1/model.neuron filter=lfs diff=lfs merge=lfs -text
4953
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_633a5bb695bb8f04bc7e/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4954
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_633a5bb695bb8f04bc7e/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
4955
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_633a5bb695bb8f04bc7e/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4956
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_633a5bb695bb8f04bc7e/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4957
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_633a5bb695bb8f04bc7e/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4958
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_64210c7d844cb8f854af/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4959
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_64210c7d844cb8f854af/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4960
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_64210c7d844cb8f854af/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4961
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_6fe7dc9962cdbf817c6b/model.neuron filter=lfs diff=lfs merge=lfs -text
4962
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_71afdf82cd71db83f05b/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4963
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_71afdf82cd71db83f05b/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4964
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_71afdf82cd71db83f05b/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4965
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_7a04700bd47382ccf70d/controlnet_0/model.neuron filter=lfs diff=lfs merge=lfs -text
4966
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_7a04700bd47382ccf70d/controlnet_1/model.neuron filter=lfs diff=lfs merge=lfs -text
4967
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_7a04700bd47382ccf70d/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4968
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_7a04700bd47382ccf70d/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4969
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_7a04700bd47382ccf70d/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4970
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_7a04700bd47382ccf70d/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4971
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_8146958fbf49c9f1e2fe/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4972
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_8146958fbf49c9f1e2fe/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
4973
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_8146958fbf49c9f1e2fe/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4974
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_8146958fbf49c9f1e2fe/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4975
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_8146958fbf49c9f1e2fe/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4976
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_88652bd9b6d7c6b771c7/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4977
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_88652bd9b6d7c6b771c7/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4978
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_88652bd9b6d7c6b771c7/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4979
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_914a39fdf2d5240e2a18/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4980
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_914a39fdf2d5240e2a18/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
4981
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_914a39fdf2d5240e2a18/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4982
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_914a39fdf2d5240e2a18/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4983
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_918d2c046c8fef1eb5fe/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4984
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_918d2c046c8fef1eb5fe/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4985
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_918d2c046c8fef1eb5fe/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4986
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_929a0f1ad1b94898582f/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4987
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_929a0f1ad1b94898582f/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4988
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_929a0f1ad1b94898582f/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4989
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9851c2d3ec41bb22bd33/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4990
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9851c2d3ec41bb22bd33/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4991
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9851c2d3ec41bb22bd33/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4992
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9b10490a6c6faca7ee0d/controlnet_0/model.neuron filter=lfs diff=lfs merge=lfs -text
4993
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9b10490a6c6faca7ee0d/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4994
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9b10490a6c6faca7ee0d/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
4995
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9b10490a6c6faca7ee0d/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
4996
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9b10490a6c6faca7ee0d/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4997
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9b10490a6c6faca7ee0d/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
4998
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9e2af0188381b43336de/model.neuron filter=lfs diff=lfs merge=lfs -text
4999
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9ffbaca0ec7a3e6e4144/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5000
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9ffbaca0ec7a3e6e4144/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5001
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_9ffbaca0ec7a3e6e4144/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5002
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_a081ee41871fedba3287/model.neuron filter=lfs diff=lfs merge=lfs -text
5003
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_a3728b20c2b54da2f938/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5004
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_a3728b20c2b54da2f938/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5005
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_a3728b20c2b54da2f938/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5006
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_a594eaf3294bb392bb3c/decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5007
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_a6a2b3ee1bf42124b53f/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5008
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_a6a2b3ee1bf42124b53f/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5009
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_a6a2b3ee1bf42124b53f/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5010
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_b5533ffa4dd2e3967489/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5011
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_b5533ffa4dd2e3967489/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5012
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_b5533ffa4dd2e3967489/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5013
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_be1cfdb7f0462b722ad7/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5014
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_be1cfdb7f0462b722ad7/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
5015
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_be1cfdb7f0462b722ad7/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5016
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_be1cfdb7f0462b722ad7/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5017
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_be1cfdb7f0462b722ad7/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5018
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_c274275c26ac0916bd57/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5019
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_c274275c26ac0916bd57/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5020
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_c274275c26ac0916bd57/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5021
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_c274275c26ac0916bd57/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5022
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_c727de806f2cb20c682c/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5023
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_c727de806f2cb20c682c/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5024
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_c727de806f2cb20c682c/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5025
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_cca040442211283fcd74/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5026
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_cca040442211283fcd74/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
5027
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_cca040442211283fcd74/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5028
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_cca040442211283fcd74/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5029
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_cca040442211283fcd74/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5030
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d349d9f345c166d4dea1/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5031
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d349d9f345c166d4dea1/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5032
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d349d9f345c166d4dea1/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5033
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d349d9f345c166d4dea1/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5034
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d6243f99d2adcd0bada5/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5035
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d6243f99d2adcd0bada5/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5036
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d6243f99d2adcd0bada5/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5037
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d9ea3ded87f7807c31d1/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5038
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d9ea3ded87f7807c31d1/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5039
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_d9ea3ded87f7807c31d1/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5040
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_dd461006bf8288f7b624/model.neuron filter=lfs diff=lfs merge=lfs -text
5041
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_e138e7001b7c9bb197a1/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5042
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_e138e7001b7c9bb197a1/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5043
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_e138e7001b7c9bb197a1/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5044
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_e138e7001b7c9bb197a1/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5045
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_e3a1b14b1d5b86b1fe3c/controlnet_0/model.neuron filter=lfs diff=lfs merge=lfs -text
5046
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_e3a1b14b1d5b86b1fe3c/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5047
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_e3a1b14b1d5b86b1fe3c/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5048
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_e3a1b14b1d5b86b1fe3c/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5049
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_e3a1b14b1d5b86b1fe3c/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5050
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_f64b18cb663010ac3340/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5051
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_f64b18cb663010ac3340/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
5052
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_f64b18cb663010ac3340/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5053
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_f64b18cb663010ac3340/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5054
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_f64b18cb663010ac3340/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5055
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fbcd61876e4f48e1119e/controlnet_0/model.neuron filter=lfs diff=lfs merge=lfs -text
5056
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fbcd61876e4f48e1119e/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5057
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fbcd61876e4f48e1119e/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
5058
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fbcd61876e4f48e1119e/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5059
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fbcd61876e4f48e1119e/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5060
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fbcd61876e4f48e1119e/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5061
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fd33833dd11008b0944d/controlnet_0/model.neuron filter=lfs diff=lfs merge=lfs -text
5062
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fd33833dd11008b0944d/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5063
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fd33833dd11008b0944d/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5064
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fd33833dd11008b0944d/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5065
+ neuronxcc-2.14.227.0+2d4f85be/MODULE_fd33833dd11008b0944d/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5066
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_1a663ed3d4415ecf05c6/model.neuron filter=lfs diff=lfs merge=lfs -text
5067
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_2503fbdba93ceb4ea7d4/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5068
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_2503fbdba93ceb4ea7d4/text_encoder_2/model.neuron filter=lfs diff=lfs merge=lfs -text
5069
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_2503fbdba93ceb4ea7d4/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
5070
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_2503fbdba93ceb4ea7d4/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5071
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_2503fbdba93ceb4ea7d4/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
5072
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_25f8e9385616f2386b3a/model.neuron filter=lfs diff=lfs merge=lfs -text
5073
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_312febc892ba028f4783/model.neuron filter=lfs diff=lfs merge=lfs -text
5074
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_96f2e7f66c8c4a2049c5/model.neuron filter=lfs diff=lfs merge=lfs -text
5075
+ neuronxcc-2.15.128.0+56dc5a86/MODULE_f4dd705de07c8a349cae/model.neuron filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/clip_text_model/hf-internal-testing/tiny-stable-diffusion-torch/0058934baf38ad450d00.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": true, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/clip_text_model/hf-internal-testing/tiny-stable-diffusion-torch/1cd8c85de423d1a4d17f.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/clip_text_model/hf-internal-testing/tiny-stable-diffusion-torch/6437b1e4638636e96aa9.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-latent-consistency/3c97e4967e6c83d09039.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 64, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 8, "num_hidden_layers": 3, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [4, 8], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 1, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 4, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 2, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": 32, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-latent-consistency/c0f11eca47e6ca3b4ca9.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 64, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 8, "num_hidden_layers": 3, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [4, 8], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 1, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 2, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": 32, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/196075233ad72b81121c.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": true, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "text_encoder_2": {"architectures": ["CLIPTextModelWithProjection"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": true, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": "text_time", "addition_embed_type_num_heads": 64, "addition_time_embed_dim": 8, "attention_head_dim": [2, 4], "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 64, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": 80, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": [1, 2], "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": true}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/5c62704e5907ab724408.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "text_encoder_2": {"architectures": ["CLIPTextModelWithProjection"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": "text_time", "addition_embed_type_num_heads": 64, "addition_time_embed_dim": 8, "attention_head_dim": [2, 4], "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 64, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": 80, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": [1, 2], "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": true}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/7560f6c90b8190fc2cc9.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "text_encoder_2": {"architectures": ["CLIPTextModelWithProjection"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": "text_time", "addition_embed_type_num_heads": 64, "addition_time_embed_dim": 8, "attention_head_dim": [2, 4], "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 64, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": 80, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": [1, 2], "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": true}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/7b816cfd640f458703b9.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "text_encoder_2": {"architectures": ["CLIPTextModelWithProjection"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": "text_time", "addition_embed_type_num_heads": 64, "addition_time_embed_dim": 8, "attention_head_dim": [2, 4], "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 64, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "all", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 4, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": 80, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": [1, 2], "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": true}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/echarlaix/tiny-random-stable-diffusion-xl/f5c5adf7a000b140312a.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "text_encoder_2": {"architectures": ["CLIPTextModelWithProjection"], "attention_dropout": 0.0, "hidden_act": "gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "output_hidden_states": true, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": "text_time", "addition_embed_type_num_heads": 64, "addition_time_embed_dim": 8, "attention_head_dim": [2, 4], "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 64, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 4, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": 80, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": [1, 2], "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": true}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/0834c9a64499cfd8014d.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 4, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/1e7de280b7fd1deb3ad9.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"controlnet_0": {"_class_name": "ControlNetModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "block_out_channels": [32, 64], "class_embed_type": null, "conditioning_channels": 3, "conditioning_embedding_out_channels": [16, 32], "controlnet_conditioning_channel_order": "rgb", "cross_attention_dim": 32, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "global_pool_conditions": false, "in_channels": 4, "layers_per_block": 2, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_encoder_hidden_size": 32, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "projection_class_embeddings_input_dim": null, "resnet_time_scale_shift": "default", "task": "semantic-segmentation", "transformer_layers_per_block": 1, "upcast_attention": false, "use_linear_projection": false}, "model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/5acab5b0d8d88e6c98af.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/6fbed240d612716ec707.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 4, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/a948b04ba5bae8fff61f.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/cd78efc0548f20d31d32.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"controlnet_0": {"_class_name": "ControlNetModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "block_out_channels": [32, 64], "class_embed_type": null, "conditioning_channels": 3, "conditioning_embedding_out_channels": [16, 32], "controlnet_conditioning_channel_order": "rgb", "cross_attention_dim": 32, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "global_pool_conditions": false, "in_channels": 4, "layers_per_block": 2, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_encoder_hidden_size": 32, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "projection_class_embeddings_input_dim": null, "resnet_time_scale_shift": "default", "task": "semantic-segmentation", "transformer_layers_per_block": 1, "upcast_attention": false, "use_linear_projection": false}, "model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/dba159d1c78cbf56d958.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/hf-internal-testing/tiny-stable-diffusion-torch/f58a4f1711d1bbd1723e.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 32, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": true, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 4, "num_hidden_layers": 5, "task": "feature-extraction", "vocab_size": 1000}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [32, 64], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 32, "cross_attention_norm": null, "down_block_types": ["DownBlock2D", "CrossAttnDownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 4, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 32, "static_num_channels": 4, "static_sequence_length": 77, "static_vae_scale_factor": 2, "static_width": 32}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["CrossAttnUpBlock2D", "UpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/stable-diffusion/timbrooks/instruct-pix2pix/20f8a909335cd43da695.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_type": "stable-diffusion", "text_encoder": {"architectures": ["CLIPTextModel"], "attention_dropout": 0.0, "dropout": 0.0, "hidden_act": "quick_gelu", "hidden_size": 768, "initializer_factor": 1.0, "initializer_range": 0.02, "intermediate_size": 3072, "layer_norm_eps": 1e-05, "max_position_embeddings": 77, "model_type": "clip_text_model", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_sequence_length": 77}, "num_attention_heads": 12, "num_hidden_layers": 12, "task": "feature-extraction", "vocab_size": 49408}, "unet": {"_class_name": "UNet2DConditionModel", "act_fn": "silu", "addition_embed_type": null, "addition_embed_type_num_heads": 64, "addition_time_embed_dim": null, "attention_head_dim": 8, "attention_type": "default", "block_out_channels": [320, 640, 1280, 1280], "center_input_sample": false, "class_embed_type": null, "class_embeddings_concat": false, "conv_in_kernel": 3, "conv_out_kernel": 3, "cross_attention_dim": 768, "cross_attention_norm": null, "down_block_types": ["CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D"], "downsample_padding": 1, "dropout": 0.0, "dual_cross_attention": false, "encoder_hid_dim": null, "encoder_hid_dim_type": null, "flip_sin_to_cos": true, "freq_shift": 0, "in_channels": 8, "layers_per_block": 2, "mid_block_only_cross_attention": null, "mid_block_scale_factor": 1, "mid_block_type": "UNetMidBlock2DCrossAttn", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_batch_size": 1, "static_height": 64, "static_num_channels": 8, "static_sequence_length": 77, "static_width": 64}, "norm_eps": 1e-05, "norm_num_groups": 32, "num_attention_heads": null, "num_class_embeds": null, "only_cross_attention": false, "out_channels": 4, "projection_class_embeddings_input_dim": null, "resnet_out_scale_factor": 1.0, "resnet_skip_time_act": false, "resnet_time_scale_shift": "default", "reverse_transformer_layers_per_block": null, "task": "semantic-segmentation", "time_cond_proj_dim": null, "time_embedding_act_fn": null, "time_embedding_dim": null, "time_embedding_type": "positional", "timestep_post_act": null, "transformer_layers_per_block": 1, "up_block_types": ["UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D"], "upcast_attention": false, "use_linear_projection": false}}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/t5/hf-internal-testing/tiny-random-t5/15d0d730286c73e4f68e.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"decoder": {"classifier_dropout": 0.0, "d_ff": 37, "d_kv": 8, "d_model": 32, "decoder_start_token_id": 0, "dense_act_fn": "relu", "dropout_rate": 0.1, "feed_forward_proj": "relu", "gradient_checkpointing": false, "initializer_factor": 0.002, "is_encoder_decoder": true, "is_gated_act": false, "layer_norm_epsilon": 1e-06, "model_type": "t5", "neuron": {"compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": true, "output_hidden_states": true, "static_batch_size": 1, "static_num_beams": 4, "static_sequence_length": 18}, "num_decoder_layers": 5, "num_heads": 4, "num_layers": 5, "relative_attention_max_distance": 128, "relative_attention_num_buckets": 8, "task": "text2text-generation", "use_cache": true, "vocab_size": 1103}, "encoder": {"classifier_dropout": 0.0, "d_ff": 37, "d_kv": 8, "d_model": 32, "decoder_start_token_id": 0, "dense_act_fn": "relu", "dropout_rate": 0.1, "feed_forward_proj": "relu", "gradient_checkpointing": false, "initializer_factor": 0.002, "is_encoder_decoder": true, "is_gated_act": false, "layer_norm_epsilon": 1e-06, "model_type": "t5", "neuron": {"compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": true, "output_hidden_states": true, "static_batch_size": 1, "static_num_beams": 4, "static_sequence_length": 18}, "num_decoder_layers": 5, "num_heads": 4, "num_layers": 5, "relative_attention_max_distance": 128, "relative_attention_num_buckets": 8, "task": "text2text-generation", "use_cache": true, "vocab_size": 1103}, "model_type": "t5"}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/facebook/wav2vec2-large-960h-lv60-self/56b22768a28f12bef9fb.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2ForCTC"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": true, "conv_dim": [512, 512, 512, 512, 512, 512, 512], "conv_kernel": [10, 3, 3, 3, 3, 2, 2], "conv_stride": [5, 2, 2, 2, 2, 2, 2], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": true, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "layer", "feat_proj_dropout": 0.1, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 4096, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 10, "mask_time_min_masks": 2, "mask_time_prob": 0.05, "model_type": "wav2vec2", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 74400, "static_batch_size": 1}, "num_adapter_layers": 3, "num_attention_heads": 16, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 16, "num_conv_pos_embeddings": 128, "num_feat_extract_layers": 7, "num_hidden_layers": 24, "num_negatives": 100, "output_hidden_size": 1024, "proj_codevector_dim": 256, "task": "audio-xvector", "tdnn_dilation": [1, 2, 3, 1, 1], "tdnn_dim": [512, 512, 512, 512, 1500], "tdnn_kernel": [5, 3, 3, 1, 1], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 512}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/facebook/wav2vec2-large-960h-lv60-self/86728684030d9f82c4d7.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2ForCTC"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": true, "conv_dim": [512, 512, 512, 512, 512, 512, 512], "conv_kernel": [10, 3, 3, 3, 3, 2, 2], "conv_stride": [5, 2, 2, 2, 2, 2, 2], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": true, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "layer", "feat_proj_dropout": 0.1, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 4096, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 10, "mask_time_min_masks": 2, "mask_time_prob": 0.05, "model_type": "wav2vec2", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 74400, "static_batch_size": 1}, "num_adapter_layers": 3, "num_attention_heads": 16, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 16, "num_conv_pos_embeddings": 128, "num_feat_extract_layers": 7, "num_hidden_layers": 24, "num_negatives": 100, "output_hidden_size": 1024, "proj_codevector_dim": 256, "task": "automatic-speech-recognition", "tdnn_dilation": [1, 2, 3, 1, 1], "tdnn_dim": [512, 512, 512, 512, 1500], "tdnn_kernel": [5, 3, 3, 1, 1], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 512}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/1f1c4a44a3fddd985756.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 74400, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "feature-extraction", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/3a64123a157e84d10501.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 74400, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "audio-xvector", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/472cb15e97bf2881387b.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "disable_fallback": false, "disable_fast_relayout": false, "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 100000, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "audio-classification", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/70bb5c3cf8828a56422c.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 74400, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "automatic-speech-recognition", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/76c77da34788c0ea062e.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "disable_fallback": false, "disable_fast_relayout": false, "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 100000, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "automatic-speech-recognition", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/92a3aa68c4cd1c5e34d3.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 4400, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "feature-extraction", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/9b9771ddbdc9ec31c7d9.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "disable_fallback": false, "disable_fast_relayout": false, "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 100000, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "audio-xvector", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/b012ed40f5e7d6b4bf0c.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "disable_fallback": false, "disable_fast_relayout": false, "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 100000, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "automatic-speech-recognition", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/ce86b7c5d629e3d3a32c.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 74400, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "audio-classification", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/db25ba0198786d9e0d25.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "disable_fallback": false, "disable_fast_relayout": false, "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 100000, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "audio-frame-classification", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/dbbb34cc758fbdfd9620.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "disable_fallback": false, "disable_fast_relayout": false, "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 100000, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "audio-xvector", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/e17307a51fd6bfd07860.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "disable_fallback": false, "disable_fast_relayout": false, "dynamic_batch_size": true, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 100000, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "audio-frame-classification", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/e58341214d08e448b289.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": null, "auto_cast_type": null, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "disable_fallback": false, "disable_fast_relayout": false, "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 100000, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "audio-classification", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.24.dev0/inference/wav2vec2/hf-internal-testing/tiny-random-Wav2Vec2Model/ed50ef1b7d06e90af218.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_dropout": 0.1, "adapter_attn_dim": null, "adapter_kernel_size": 3, "adapter_stride": 2, "add_adapter": false, "apply_spec_augment": true, "architectures": ["Wav2Vec2Model"], "attention_dropout": 0.1, "classifier_proj_size": 256, "codevector_dim": 256, "contrastive_logits_temperature": 0.1, "conv_bias": false, "conv_dim": [32, 32, 32], "conv_kernel": [8, 8, 8], "conv_stride": [4, 4, 4], "ctc_loss_reduction": "sum", "ctc_zero_infinity": false, "diversity_loss_weight": 0.1, "do_stable_layer_norm": false, "feat_extract_activation": "gelu", "feat_extract_dropout": 0.0, "feat_extract_norm": "group", "feat_proj_dropout": 0.0, "feat_quantizer_dropout": 0.0, "final_dropout": 0.1, "hidden_act": "gelu", "hidden_dropout": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 20, "layer_norm_eps": 1e-05, "layerdrop": 0.1, "mask_feature_length": 10, "mask_feature_min_masks": 0, "mask_feature_prob": 0.0, "mask_time_length": 2, "mask_time_min_masks": 2, "mask_time_prob": 0.5, "model_type": "wav2vec2", "neuron": {"auto_cast": "matmul", "auto_cast_type": "bf16", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "dynamic_batch_size": false, "inline_weights_to_neff": false, "optlevel": "2", "output_attentions": false, "output_hidden_states": false, "static_audio_sequence_length": 74400, "static_batch_size": 1}, "num_adapter_layers": 1, "num_attention_heads": 2, "num_codevector_groups": 2, "num_codevectors_per_group": 320, "num_conv_pos_embedding_groups": 2, "num_conv_pos_embeddings": 16, "num_feat_extract_layers": 3, "num_hidden_layers": 4, "num_negatives": 100, "output_hidden_size": 16, "proj_codevector_dim": 256, "task": "audio-frame-classification", "tdnn_dilation": [1, 2], "tdnn_dim": [32, 32], "tdnn_kernel": [5, 3], "use_weighted_layer_sum": false, "vocab_size": 32, "xvector_output_dim": 32}
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_center_crop",
8
+ "crop_size",
9
+ "do_rescale",
10
+ "rescale_factor",
11
+ "do_normalize",
12
+ "image_mean",
13
+ "image_std",
14
+ "do_convert_rgb",
15
+ "return_tensors",
16
+ "data_format",
17
+ "input_data_format"
18
+ ],
19
+ "crop_size": {
20
+ "height": 224,
21
+ "width": 224
22
+ },
23
+ "do_center_crop": true,
24
+ "do_convert_rgb": true,
25
+ "do_normalize": true,
26
+ "do_rescale": true,
27
+ "do_resize": true,
28
+ "image_mean": [
29
+ 0.48145466,
30
+ 0.4578275,
31
+ 0.40821073
32
+ ],
33
+ "image_processor_type": "CLIPImageProcessor",
34
+ "image_std": [
35
+ 0.26862954,
36
+ 0.26130258,
37
+ 0.27577711
38
+ ],
39
+ "resample": 3,
40
+ "rescale_factor": 0.00392156862745098,
41
+ "size": {
42
+ "shortest_edge": 224
43
+ }
44
+ }
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/model_index.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.29.1",
4
+ "_name_or_path": "hf-internal-testing/tiny-stable-diffusion-torch",
5
+ "feature_extractor": [
6
+ "transformers",
7
+ "CLIPImageProcessor"
8
+ ],
9
+ "image_encoder": [
10
+ null,
11
+ null
12
+ ],
13
+ "requires_safety_checker": true,
14
+ "safety_checker": [
15
+ null,
16
+ null
17
+ ],
18
+ "scheduler": [
19
+ "diffusers",
20
+ "PNDMScheduler"
21
+ ],
22
+ "text_encoder": [
23
+ "transformers",
24
+ "CLIPTextModel"
25
+ ],
26
+ "tokenizer": [
27
+ "transformers",
28
+ "CLIPTokenizer"
29
+ ],
30
+ "unet": [
31
+ "diffusers",
32
+ "UNet2DConditionModel"
33
+ ],
34
+ "vae": [
35
+ "diffusers",
36
+ "AutoencoderKL"
37
+ ]
38
+ }
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/scheduler/scheduler_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.29.1",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
+ "set_alpha_to_one": false,
11
+ "skip_prk_steps": true,
12
+ "steps_offset": 1,
13
+ "timestep_spacing": "leading",
14
+ "trained_betas": null
15
+ }
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/text_encoder/config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/home/ubuntu/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-torch/snapshots/a88cdfbd91f96ec7f61eb7484b652ff0f4ee701d/text_encoder",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 32,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 37,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "neuron": {
19
+ "auto_cast": "matmul",
20
+ "auto_cast_type": "bf16",
21
+ "compiler_type": "neuronx-cc",
22
+ "compiler_version": "2.13.66.0+6dfecc895",
23
+ "dynamic_batch_size": false,
24
+ "inline_weights_to_neff": true,
25
+ "input_names": [
26
+ "input_ids"
27
+ ],
28
+ "model_type": "clip-text-model",
29
+ "optlevel": "2",
30
+ "output_attentions": false,
31
+ "output_hidden_states": true,
32
+ "output_names": [
33
+ "last_hidden_state",
34
+ "pooler_output"
35
+ ],
36
+ "static_batch_size": 1,
37
+ "static_sequence_length": 77
38
+ },
39
+ "num_attention_heads": 4,
40
+ "num_hidden_layers": 5,
41
+ "pad_token_id": 1,
42
+ "projection_dim": 512,
43
+ "task": "feature-extraction",
44
+ "torch_dtype": "float32",
45
+ "torchscript": true,
46
+ "transformers_version": "4.41.1",
47
+ "vocab_size": 1000
48
+ }
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/text_encoder/model.neuron ADDED
Binary file (877 kB). View file
 
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/tokenizer/merges.txt ADDED
@@ -0,0 +1,647 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #version: 0.2
2
+ Ġ t
3
+ Ġt h
4
+ Ġ a
5
+ Ġth e</w>
6
+ i n
7
+ Ġ o
8
+ Ġ ,</w>
9
+ Ġ s
10
+ e d</w>
11
+ Ġ w
12
+ e r
13
+ Ġ .</w>
14
+ Ġ i
15
+ r e
16
+ Ġ c
17
+ n d</w>
18
+ Ġ f
19
+ Ġ b
20
+ a t
21
+ Ġo f</w>
22
+ e r</w>
23
+ e n
24
+ a r
25
+ o r
26
+ i t
27
+ Ġ p
28
+ Ġ h
29
+ Ġa nd</w>
30
+ o n
31
+ in g</w>
32
+ a n
33
+ r o
34
+ Ġ m
35
+ Ġ d
36
+ e s</w>
37
+ Ġi n</w>
38
+ o n</w>
39
+ Ġt o</w>
40
+ o u
41
+ i s
42
+ Ġ a</w>
43
+ i c
44
+ Ġ T
45
+ a l
46
+ Ġ l
47
+ Ġ =</w>
48
+ Ġ re
49
+ Ġ "</w>
50
+ e s
51
+ Ġ S
52
+ a s</w>
53
+ a l</w>
54
+ i l
55
+ e l
56
+ i on</w>
57
+ Ġ A
58
+ Ġ C
59
+ Ġ 1
60
+ Ġ Ċ</w>
61
+ u r
62
+ ĠT h
63
+ Ġ n
64
+ a s
65
+ Ġ @
66
+ e c
67
+ o m
68
+ a c
69
+ Ġ e
70
+ Ġw as</w>
71
+ Ġ M
72
+ o r</w>
73
+ a n</w>
74
+ a m
75
+ e n</w>
76
+ o l
77
+ Ġ in
78
+ Ġ g
79
+ Ġ '</w>
80
+ Ġ B
81
+ l y</w>
82
+ a t</w>
83
+ i v
84
+ t s</w>
85
+ ĠTh e</w>
86
+ u s
87
+ - @</w>
88
+ Ġ@ -@</w>
89
+ i s</w>
90
+ Ġ I
91
+ Ġw h
92
+ i g
93
+ Ġ H
94
+ Ġs t
95
+ o s
96
+ u n
97
+ t h
98
+ Ġ P
99
+ Ġw it
100
+ Ġth at</w>
101
+ i r
102
+ Ġa s</w>
103
+ e m
104
+ Ġo n</w>
105
+ r a
106
+ Ġf or</w>
107
+ Ġ R
108
+ e t
109
+ o w
110
+ Ġ 2
111
+ i d
112
+ Ġ D
113
+ l e</w>
114
+ Ġwit h</w>
115
+ l a
116
+ en t</w>
117
+ i m
118
+ Ġ F
119
+ e a
120
+ i on
121
+ Ġb y</w>
122
+ Ġ )</w>
123
+ Ġ (</w>
124
+ Ġa l
125
+ Ġc on
126
+ en t
127
+ Ġ W
128
+ Ġi s</w>
129
+ er e</w>
130
+ Ġ G
131
+ Ġ N
132
+ Ġ L
133
+ Ġh a
134
+ er s</w>
135
+ r i
136
+ t h</w>
137
+ t ed</w>
138
+ u c
139
+ Ġ J
140
+ Ġ1 9
141
+ e v
142
+ u l
143
+ Ġ v
144
+ c e</w>
145
+ at ion</w>
146
+ ro m</w>
147
+ Ġb e
148
+ Ġ E
149
+ i n</w>
150
+ Ġth e
151
+ Ġf rom</w>
152
+ Ġ O
153
+ t er</w>
154
+ Ġp ro
155
+ Ġa r
156
+ a d
157
+ Ġc om
158
+ i c</w>
159
+ a g
160
+ Ġh is</w>
161
+ Ġs h
162
+ Ġa t</w>
163
+ o v
164
+ i es</w>
165
+ o o
166
+ p p
167
+ s t
168
+ c h
169
+ Ġ r
170
+ Ġ2 0
171
+ a y</w>
172
+ i f
173
+ Ġw ere</w>
174
+ Ġc h
175
+ u t</w>
176
+ s t</w>
177
+ u t
178
+ d s</w>
179
+ o p
180
+ u m
181
+ Ġi t</w>
182
+ o c
183
+ t er
184
+ l e
185
+ ig h
186
+ u d
187
+ Ġe x
188
+ ion s</w>
189
+ at e</w>
190
+ it y</w>
191
+ at ed</w>
192
+ Ġ un
193
+ e p
194
+ q u
195
+ Ġn o
196
+ Ġ K
197
+ iv e</w>
198
+ is t
199
+ Ġo n
200
+ am e</w>
201
+ ou n
202
+ i r</w>
203
+ a b
204
+ Ġ â
205
+ in g
206
+ Ġh e</w>
207
+ l d</w>
208
+ u g
209
+ ic h</w>
210
+ Ġa n</w>
211
+ e d
212
+ Ġ k
213
+ Ġâ Ģ
214
+ Ġha d</w>
215
+ v e</w>
216
+ a in
217
+ Ġs e
218
+ t ion</w>
219
+ or e</w>
220
+ re s
221
+ Ġwh ich</w>
222
+ ĠI n</w>
223
+ o d
224
+ th er</w>
225
+ a k
226
+ Ġs p
227
+ a r</w>
228
+ Ġ y
229
+ ĠC h
230
+ on g</w>
231
+ Ġa c
232
+ es t</w>
233
+ Ġ U
234
+ a p
235
+ f f
236
+ al ly</w>
237
+ r it
238
+ ĠS t
239
+ u b
240
+ g e</w>
241
+ b er</w>
242
+ e t</w>
243
+ Ġb e</w>
244
+ e ar
245
+ Ġre c
246
+ er s
247
+ Ġf ir
248
+ o t
249
+ Ġar e</w>
250
+ Ġa n
251
+ c h</w>
252
+ o g
253
+ i a</w>
254
+ es t
255
+ in e</w>
256
+ il l
257
+ an d
258
+ e l</w>
259
+ ar y</w>
260
+ e w</w>
261
+ i d</w>
262
+ Ġf or
263
+ Ġ ;</w>
264
+ Ġcom p
265
+ Ġ V
266
+ Ġin c
267
+ t r
268
+ Ġ20 0
269
+ Ġthe ir</w>
270
+ u s</w>
271
+ Ġb ut</w>
272
+ r an
273
+ ic al</w>
274
+ Ġfir st</w>
275
+ Ġd e
276
+ Ġin t
277
+ Ġ ro
278
+ s o</w>
279
+ ĠâĢ ĵ</w>
280
+ Ġno t</w>
281
+ d ing</w>
282
+ f ter</w>
283
+ ur e</w>
284
+ Ġp ar
285
+ Ġ :</w>
286
+ i an</w>
287
+ Ġt w
288
+ ou ld</w>
289
+ Ġal so</w>
290
+ Ġi ts</w>
291
+ Ġw or
292
+ u m</w>
293
+ Ġo r</w>
294
+ os t</w>
295
+ 0 0</w>
296
+ ou r
297
+ ar d</w>
298
+ Ġre s
299
+ m p
300
+ u e</w>
301
+ Ġa b
302
+ is h</w>
303
+ Ġcon t
304
+ Ġa d
305
+ ow n</w>
306
+ al l</w>
307
+ ou g
308
+ Ġh er</w>
309
+ as t</w>
310
+ Ġ en
311
+ om e</w>
312
+ al l
313
+ d ed</w>
314
+ o w</w>
315
+ Ġha ve</w>
316
+ Ġ us
317
+ ea r</w>
318
+ ac k</w>
319
+ d uc
320
+ i al</w>
321
+ s s
322
+ en ts</w>
323
+ a in</w>
324
+ t ing</w>
325
+ Ġon e</w>
326
+ es s
327
+ Ġh as</w>
328
+ igh t</w>
329
+ a v
330
+ Ġe v
331
+ ou t</w>
332
+ a y
333
+ en ce</w>
334
+ Ġbe en</w>
335
+ e w
336
+ Ġtw o</w>
337
+ Ġc l
338
+ d er</w>
339
+ im e</w>
340
+ k s</w>
341
+ es s</w>
342
+ is h
343
+ . @</w>
344
+ Ġ@ .@</w>
345
+ Ġp la
346
+ Ġp l
347
+ Ġo r
348
+ u p</w>
349
+ m ent</w>
350
+ ur ing</w>
351
+ ol l
352
+ ĠI n
353
+ Ġth is</w>
354
+ Ġb ec
355
+ Ġcom m
356
+ Ġd is
357
+ at er</w>
358
+ ag e</w>
359
+ Ġa pp
360
+ ou s</w>
361
+ e y</w>
362
+ i l</w>
363
+ p er
364
+ ĠA l
365
+ ion al</w>
366
+ l ud
367
+ el y</w>
368
+ t t
369
+ il e</w>
370
+ i z
371
+ Ġ j
372
+ Ġwh o</w>
373
+ Ġa g
374
+ i b
375
+ Ġthe y</w>
376
+ f or
377
+ Ġo v
378
+ at h
379
+ e g
380
+ Ġs c
381
+ i p
382
+ Ġ20 1
383
+ Ġ 3
384
+ Ġp er
385
+ or y</w>
386
+ Ġd es
387
+ id e</w>
388
+ Ġs er
389
+ s e</w>
390
+ ĠH e</w>
391
+ la nd</w>
392
+ at ions</w>
393
+ r ic
394
+ i t</w>
395
+ re s</w>
396
+ er ed</w>
397
+ Ġp re
398
+ ĠS h
399
+ an ce</w>
400
+ or t</w>
401
+ an t</w>
402
+ , @</w>
403
+ Ġ@ ,@</w>
404
+ el l</w>
405
+ Ġ Y
406
+ n ed</w>
407
+ el l
408
+ it e</w>
409
+ Ġinc lud
410
+ Ġre p
411
+ Ġa fter</w>
412
+ Ġs uc
413
+ re e</w>
414
+ an y</w>
415
+ i m</w>
416
+ or t
417
+ Ġ1 8
418
+ Ġs u
419
+ ad e</w>
420
+ ou r</w>
421
+ ĠU n
422
+ ĠI t</w>
423
+ i k
424
+ ĠM ar
425
+ em ber</w>
426
+ Ġ 1</w>
427
+ e en</w>
428
+ a nd</w>
429
+ Ġs ec
430
+ ic e</w>
431
+ Ġt ime</w>
432
+ ĠA n
433
+ Ġint o</w>
434
+ Ġf in
435
+ Ġo ther</w>
436
+ Ġa tt
437
+ il l</w>
438
+ re n
439
+ ac h
440
+ as s
441
+ er al</w>
442
+ es e</w>
443
+ s h
444
+ al s</w>
445
+ it ion</w>
446
+ oug h</w>
447
+ l es</w>
448
+ am p
449
+ Ġw ould</w>
450
+ Ġm ore</w>
451
+ ro ug
452
+ ri b
453
+ er y</w>
454
+ ac e</w>
455
+ Ġ A</w>
456
+ Ġpla y
457
+ it ed</w>
458
+ k ed</w>
459
+ is t</w>
460
+ i ed</w>
461
+ Ġ 2</w>
462
+ as ed</w>
463
+ ing s</w>
464
+ an g
465
+ a m</w>
466
+ i p</w>
467
+ Ġb o
468
+ ab le</w>
469
+ t y</w>
470
+ Ġch ar
471
+ Ġc ent
472
+ et w
473
+ at es</w>
474
+ ro p
475
+ Ġ I</w>
476
+ u nd</w>
477
+ ĠA m
478
+ c es</w>
479
+ o in
480
+ Ġin ter
481
+ u p
482
+ c t
483
+ on e</w>
484
+ Ġt ra
485
+ an t
486
+ ec t
487
+ Ġal l</w>
488
+ e f
489
+ Ġcon s
490
+ ub l
491
+ n ing</w>
492
+ an s</w>
493
+ Ġf e
494
+ us t</w>
495
+ Ġ 0
496
+ Ġre m
497
+ as e</w>
498
+ on g
499
+ Ġwh en</w>
500
+ e b
501
+ ĠW h
502
+ Ġe ar
503
+ ev er</w>
504
+ Ġov er</w>
505
+ Ġk n
506
+ a us
507
+ Ġp os
508
+ a d</w>
509
+ er m
510
+ Ġsh e</w>
511
+ Ġ ra
512
+ Ġd uring</w>
513
+ as on</w>
514
+ v i
515
+ Ġex p
516
+ Ġl ea
517
+ Ġ el
518
+ Ġ 4
519
+ Ġon ly</w>
520
+ o nd</w>
521
+ Ġd ec
522
+ Ġac c
523
+ Ġo ff
524
+ is s
525
+ Ġf l
526
+ ĠE n
527
+ o t</w>
528
+ en s
529
+ os e</w>
530
+ ak e</w>
531
+ o m</w>
532
+ Ġs ev
533
+ ac h</w>
534
+ etw een</w>
535
+ er n
536
+ Ġ 3</w>
537
+ Ġp r
538
+ Ġg ro
539
+ r uc
540
+ Ġd i
541
+ Ġ19 9
542
+ ĠA r
543
+ Ġg ame</w>
544
+ Ġh im</w>
545
+ oo k</w>
546
+ Ġ up</w>
547
+ Ġab out</w>
548
+ Ġre l
549
+ for m
550
+ Ġth ree</w>
551
+ at t
552
+ ĠC om
553
+ Ġs a
554
+ ear s</w>
555
+ Ġ 5
556
+ r y</w>
557
+ Ġi mp
558
+ Ġm ost</w>
559
+ f er
560
+ Ġp res
561
+ Ġf il
562
+ Ġb etween</w>
563
+ Ġbe g
564
+ p h
565
+ or s</w>
566
+ Ġth an</w>
567
+ Ġrec or
568
+ o b
569
+ er ic
570
+ at ing</w>
571
+ Ġth roug
572
+ k ing</w>
573
+ Ġo ut</w>
574
+ Ġn um
575
+ oo d</w>
576
+ oll ow
577
+ ac t
578
+ u il
579
+ Ġc re
580
+ ol og
581
+ at ional</w>
582
+ Ġpro duc
583
+ Ġwh ile</w>
584
+ Ġl ater</w>
585
+ Ġw rit
586
+ e x
587
+ Ġst ar
588
+ Ġsp ec
589
+ e e
590
+ ish ed</w>
591
+ Ġre g
592
+ is ion</w>
593
+ ou th</w>
594
+ Ġre le
595
+ Ġa ss
596
+ Ġse ason</w>
597
+ Ġm ade</w>
598
+ il y</w>
599
+ r u
600
+ o y
601
+ t ur
602
+ t e</w>
603
+ Ġ qu
604
+ Ġm ov
605
+ ur y</w>
606
+ ĠAm eric
607
+ em ent</w>
608
+ c c
609
+ ou nd</w>
610
+ Ġl ar
611
+ Ġfor m
612
+ ec t</w>
613
+ Ġde f
614
+ Ġm us
615
+ ĠP ar
616
+ Ġm e
617
+ Ġs ub
618
+ w ay</w>
619
+ o p</w>
620
+ o h
621
+ el d</w>
622
+ i e</w>
623
+ em p
624
+ am es</w>
625
+ er n</w>
626
+ Ġn or
627
+ iv ed</w>
628
+ ev el
629
+ Ġsuc h</w>
630
+ ar ds</w>
631
+ Ġin d
632
+ ik e</w>
633
+ Ġg en
634
+ er t
635
+ Ġy ear</w>
636
+ Ġus ed</w>
637
+ Ġn ew</w>
638
+ Ġ 5</w>
639
+ Ġal b
640
+ s p
641
+ y p
642
+ Ġwit h
643
+ Ġwh ere</w>
644
+ ic s</w>
645
+ ĠTh is</w>
646
+ Ġthe m</w>
647
+ w n</w>
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<|endoftext|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ }
20
+ },
21
+ "bos_token": "<|startoftext|>",
22
+ "clean_up_tokenization_spaces": true,
23
+ "do_lower_case": true,
24
+ "eos_token": "<|endoftext|>",
25
+ "errors": "replace",
26
+ "model_max_length": 77,
27
+ "pad_token": "<|endoftext|>",
28
+ "tokenizer_class": "CLIPTokenizer",
29
+ "unk_token": "<|endoftext|>"
30
+ }
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/tokenizer/vocab.json ADDED
@@ -0,0 +1,1002 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "!": 2,
3
+ "!</w>": 345,
4
+ "\"": 3,
5
+ "\"</w>": 344,
6
+ "#": 4,
7
+ "#</w>": 325,
8
+ "$": 5,
9
+ "$</w>": 348,
10
+ "%": 6,
11
+ "%</w>": 351,
12
+ "&": 7,
13
+ "&</w>": 352,
14
+ "'": 8,
15
+ "'</w>": 296,
16
+ "(": 9,
17
+ "(</w>": 318,
18
+ ")": 10,
19
+ ")</w>": 330,
20
+ "*": 11,
21
+ "*</w>": 327,
22
+ "+": 12,
23
+ "+</w>": 341,
24
+ ",": 13,
25
+ ",</w>": 279,
26
+ ",@</w>": 754,
27
+ "-": 14,
28
+ "-</w>": 276,
29
+ "-@</w>": 439,
30
+ ".": 15,
31
+ ".</w>": 253,
32
+ ".@</w>": 695,
33
+ "/": 16,
34
+ "/</w>": 350,
35
+ "0": 17,
36
+ "00</w>": 647,
37
+ "0</w>": 216,
38
+ "1": 18,
39
+ "1</w>": 222,
40
+ "2": 19,
41
+ "2</w>": 231,
42
+ "3": 20,
43
+ "3</w>": 243,
44
+ "4": 21,
45
+ "4</w>": 233,
46
+ "5": 22,
47
+ "5</w>": 240,
48
+ "6": 23,
49
+ "6</w>": 226,
50
+ "7": 24,
51
+ "7</w>": 215,
52
+ "8": 25,
53
+ "8</w>": 236,
54
+ "9": 26,
55
+ "9</w>": 242,
56
+ ":": 27,
57
+ ":</w>": 353,
58
+ ";": 28,
59
+ ";</w>": 317,
60
+ "<": 29,
61
+ "<</w>": 340,
62
+ "<|endoftext|>": 1,
63
+ "<|startoftext|>": 0,
64
+ "=": 30,
65
+ "=</w>": 342,
66
+ ">": 31,
67
+ "></w>": 300,
68
+ "?": 32,
69
+ "?</w>": 346,
70
+ "@": 33,
71
+ "@</w>": 320,
72
+ "A": 34,
73
+ "A</w>": 227,
74
+ "B": 35,
75
+ "B</w>": 258,
76
+ "C": 36,
77
+ "C</w>": 239,
78
+ "D": 37,
79
+ "D</w>": 255,
80
+ "E": 38,
81
+ "E</w>": 246,
82
+ "F": 39,
83
+ "F</w>": 213,
84
+ "G": 40,
85
+ "G</w>": 283,
86
+ "H": 41,
87
+ "H</w>": 219,
88
+ "I": 42,
89
+ "I</w>": 237,
90
+ "J": 43,
91
+ "J</w>": 251,
92
+ "K": 44,
93
+ "K</w>": 254,
94
+ "L": 45,
95
+ "L</w>": 218,
96
+ "M": 46,
97
+ "M</w>": 234,
98
+ "N": 47,
99
+ "N</w>": 238,
100
+ "O": 48,
101
+ "O</w>": 265,
102
+ "P": 49,
103
+ "P</w>": 245,
104
+ "Q": 50,
105
+ "Q</w>": 309,
106
+ "R": 51,
107
+ "R</w>": 264,
108
+ "S": 52,
109
+ "S</w>": 230,
110
+ "T": 53,
111
+ "T</w>": 235,
112
+ "U": 54,
113
+ "U</w>": 268,
114
+ "V": 55,
115
+ "V</w>": 248,
116
+ "W": 56,
117
+ "W</w>": 274,
118
+ "X": 57,
119
+ "X</w>": 263,
120
+ "Y": 58,
121
+ "Y</w>": 310,
122
+ "Z": 59,
123
+ "Z</w>": 207,
124
+ "[": 60,
125
+ "[</w>": 270,
126
+ "\\": 61,
127
+ "\\</w>": 338,
128
+ "]": 62,
129
+ "]</w>": 289,
130
+ "^": 63,
131
+ "^</w>": 331,
132
+ "_": 64,
133
+ "_</w>": 334,
134
+ "`": 65,
135
+ "`</w>": 347,
136
+ "a": 66,
137
+ "a</w>": 197,
138
+ "ab": 555,
139
+ "able</w>": 820,
140
+ "ac": 420,
141
+ "ace</w>": 806,
142
+ "ach": 791,
143
+ "ach</w>": 885,
144
+ "ack</w>": 670,
145
+ "act": 929,
146
+ "ad": 508,
147
+ "ad</w>": 860,
148
+ "ade</w>": 771,
149
+ "ag": 511,
150
+ "age</w>": 710,
151
+ "ain": 568,
152
+ "ain</w>": 675,
153
+ "ak": 577,
154
+ "ake</w>": 882,
155
+ "al": 397,
156
+ "al</w>": 405,
157
+ "all": 664,
158
+ "all</w>": 658,
159
+ "ally</w>": 588,
160
+ "als</w>": 796,
161
+ "am": 426,
162
+ "am</w>": 817,
163
+ "ame</w>": 552,
164
+ "ames</w>": 976,
165
+ "amp": 800,
166
+ "an": 384,
167
+ "an</w>": 425,
168
+ "ance</w>": 751,
169
+ "and": 609,
170
+ "and</w>": 780,
171
+ "ang": 816,
172
+ "ans</w>": 844,
173
+ "ant": 837,
174
+ "ant</w>": 753,
175
+ "any</w>": 766,
176
+ "ap": 586,
177
+ "ar": 376,
178
+ "ar</w>": 579,
179
+ "ard</w>": 649,
180
+ "ards</w>": 982,
181
+ "ary</w>": 611,
182
+ "as": 416,
183
+ "as</w>": 404,
184
+ "ase</w>": 849,
185
+ "ased</w>": 814,
186
+ "ason</w>": 865,
187
+ "ass": 792,
188
+ "ast</w>": 661,
189
+ "at": 372,
190
+ "at</w>": 434,
191
+ "ate</w>": 541,
192
+ "ated</w>": 543,
193
+ "ater</w>": 709,
194
+ "ates</w>": 825,
195
+ "ath": 730,
196
+ "ating</w>": 922,
197
+ "ation</w>": 497,
198
+ "ational</w>": 933,
199
+ "ations</w>": 744,
200
+ "att": 903,
201
+ "aus": 858,
202
+ "av": 681,
203
+ "ay": 684,
204
+ "ay</w>": 523,
205
+ "b": 67,
206
+ "b</w>": 212,
207
+ "ber</w>": 593,
208
+ "c": 68,
209
+ "c</w>": 224,
210
+ "cc": 960,
211
+ "ce</w>": 496,
212
+ "ces</w>": 830,
213
+ "ch": 520,
214
+ "ch</w>": 603,
215
+ "ct": 834,
216
+ "d": 69,
217
+ "d</w>": 196,
218
+ "ded</w>": 665,
219
+ "der</w>": 690,
220
+ "ding</w>": 633,
221
+ "ds</w>": 530,
222
+ "duc": 671,
223
+ "e": 70,
224
+ "e</w>": 195,
225
+ "ea": 471,
226
+ "ear": 596,
227
+ "ear</w>": 669,
228
+ "ears</w>": 906,
229
+ "eb": 852,
230
+ "ec": 418,
231
+ "ect": 838,
232
+ "ect</w>": 964,
233
+ "ed": 563,
234
+ "ed</w>": 362,
235
+ "ee": 941,
236
+ "een</w>": 779,
237
+ "ef": 840,
238
+ "eg": 731,
239
+ "el": 407,
240
+ "el</w>": 610,
241
+ "eld</w>": 973,
242
+ "ell": 759,
243
+ "ell</w>": 756,
244
+ "ely</w>": 719,
245
+ "em": 455,
246
+ "ember</w>": 777,
247
+ "ement</w>": 959,
248
+ "emp": 975,
249
+ "en": 375,
250
+ "en</w>": 427,
251
+ "ence</w>": 685,
252
+ "ens": 880,
253
+ "ent": 478,
254
+ "ent</w>": 468,
255
+ "ents</w>": 674,
256
+ "ep": 545,
257
+ "er": 364,
258
+ "er</w>": 374,
259
+ "eral</w>": 793,
260
+ "ere</w>": 481,
261
+ "ered</w>": 748,
262
+ "eric": 921,
263
+ "erm": 861,
264
+ "ern": 887,
265
+ "ern</w>": 977,
266
+ "ers": 598,
267
+ "ers</w>": 486,
268
+ "ert": 986,
269
+ "ery</w>": 805,
270
+ "es": 402,
271
+ "es</w>": 388,
272
+ "ese</w>": 794,
273
+ "ess": 678,
274
+ "ess</w>": 693,
275
+ "est": 606,
276
+ "est</w>": 584,
277
+ "et": 460,
278
+ "et</w>": 594,
279
+ "etw": 824,
280
+ "etween</w>": 886,
281
+ "ev": 493,
282
+ "evel": 980,
283
+ "ever</w>": 855,
284
+ "ew": 687,
285
+ "ew</w>": 612,
286
+ "ex": 938,
287
+ "ey</w>": 713,
288
+ "f": 71,
289
+ "f</w>": 209,
290
+ "fer": 911,
291
+ "ff": 587,
292
+ "for": 728,
293
+ "form": 901,
294
+ "fter</w>": 634,
295
+ "g": 72,
296
+ "g</w>": 214,
297
+ "ge</w>": 592,
298
+ "h": 73,
299
+ "h</w>": 203,
300
+ "i": 74,
301
+ "i</w>": 205,
302
+ "ia</w>": 605,
303
+ "ial</w>": 672,
304
+ "ian</w>": 638,
305
+ "ib": 726,
306
+ "ic": 395,
307
+ "ic</w>": 510,
308
+ "ical</w>": 625,
309
+ "ice</w>": 782,
310
+ "ich</w>": 561,
311
+ "ics</w>": 996,
312
+ "id": 463,
313
+ "id</w>": 613,
314
+ "ide</w>": 739,
315
+ "ie</w>": 974,
316
+ "ied</w>": 812,
317
+ "ies</w>": 516,
318
+ "if": 524,
319
+ "ig": 444,
320
+ "igh": 537,
321
+ "ight</w>": 680,
322
+ "ik": 775,
323
+ "ike</w>": 984,
324
+ "il": 406,
325
+ "il</w>": 714,
326
+ "ile</w>": 721,
327
+ "ill": 608,
328
+ "ill</w>": 789,
329
+ "ily</w>": 950,
330
+ "im": 469,
331
+ "im</w>": 767,
332
+ "ime</w>": 691,
333
+ "in": 358,
334
+ "in</w>": 501,
335
+ "ine</w>": 607,
336
+ "ing": 557,
337
+ "ing</w>": 383,
338
+ "ings</w>": 815,
339
+ "ion": 472,
340
+ "ion</w>": 408,
341
+ "ional</w>": 717,
342
+ "ions</w>": 540,
343
+ "ip": 733,
344
+ "ip</w>": 818,
345
+ "ir": 453,
346
+ "ir</w>": 554,
347
+ "is": 393,
348
+ "is</w>": 441,
349
+ "ish": 694,
350
+ "ish</w>": 654,
351
+ "ished</w>": 942,
352
+ "ision</w>": 944,
353
+ "iss": 876,
354
+ "ist": 550,
355
+ "ist</w>": 811,
356
+ "it": 378,
357
+ "it</w>": 746,
358
+ "ite</w>": 760,
359
+ "ited</w>": 809,
360
+ "ition</w>": 797,
361
+ "ity</w>": 542,
362
+ "iv": 435,
363
+ "ive</w>": 549,
364
+ "ived</w>": 979,
365
+ "iz": 722,
366
+ "j": 75,
367
+ "j</w>": 288,
368
+ "k": 76,
369
+ "k</w>": 210,
370
+ "ked</w>": 810,
371
+ "king</w>": 924,
372
+ "ks</w>": 692,
373
+ "l": 77,
374
+ "l</w>": 201,
375
+ "la": 467,
376
+ "land</w>": 743,
377
+ "ld</w>": 559,
378
+ "le": 536,
379
+ "le</w>": 465,
380
+ "les</w>": 799,
381
+ "lud": 718,
382
+ "ly</w>": 433,
383
+ "m": 78,
384
+ "m</w>": 202,
385
+ "ment</w>": 701,
386
+ "mp": 651,
387
+ "n": 79,
388
+ "n</w>": 199,
389
+ "nd</w>": 369,
390
+ "ned</w>": 758,
391
+ "ning</w>": 843,
392
+ "o": 80,
393
+ "o</w>": 198,
394
+ "ob": 920,
395
+ "oc": 534,
396
+ "od": 575,
397
+ "og": 604,
398
+ "oh": 972,
399
+ "oin": 831,
400
+ "ol": 428,
401
+ "oll": 703,
402
+ "ollow": 928,
403
+ "olog": 932,
404
+ "om": 419,
405
+ "om</w>": 883,
406
+ "ome</w>": 663,
407
+ "on": 382,
408
+ "on</w>": 390,
409
+ "ond</w>": 872,
410
+ "one</w>": 835,
411
+ "ong": 850,
412
+ "ong</w>": 582,
413
+ "oo": 517,
414
+ "ood</w>": 927,
415
+ "ook</w>": 897,
416
+ "op": 531,
417
+ "op</w>": 971,
418
+ "or": 377,
419
+ "or</w>": 424,
420
+ "ore</w>": 571,
421
+ "ors</w>": 917,
422
+ "ort": 768,
423
+ "ort</w>": 752,
424
+ "ory</w>": 737,
425
+ "os": 447,
426
+ "ose</w>": 881,
427
+ "ost</w>": 646,
428
+ "ot": 600,
429
+ "ot</w>": 879,
430
+ "ou": 392,
431
+ "oug": 659,
432
+ "ough</w>": 798,
433
+ "ould</w>": 640,
434
+ "oun": 553,
435
+ "ound</w>": 961,
436
+ "our": 648,
437
+ "our</w>": 772,
438
+ "ous</w>": 712,
439
+ "out</w>": 683,
440
+ "outh</w>": 945,
441
+ "ov": 515,
442
+ "ow": 461,
443
+ "ow</w>": 666,
444
+ "own</w>": 657,
445
+ "oy": 952,
446
+ "p": 81,
447
+ "p</w>": 217,
448
+ "per": 715,
449
+ "ph": 916,
450
+ "pp": 518,
451
+ "q": 82,
452
+ "q</w>": 280,
453
+ "qu": 546,
454
+ "r": 83,
455
+ "r</w>": 204,
456
+ "ra": 457,
457
+ "ran": 624,
458
+ "re": 367,
459
+ "ree</w>": 765,
460
+ "ren": 790,
461
+ "res": 572,
462
+ "res</w>": 747,
463
+ "ri": 487,
464
+ "rib": 804,
465
+ "ric": 745,
466
+ "rit": 589,
467
+ "ro": 385,
468
+ "rom</w>": 498,
469
+ "rop": 826,
470
+ "roug": 803,
471
+ "ru": 951,
472
+ "ruc": 891,
473
+ "ry</w>": 908,
474
+ "s": 84,
475
+ "s</w>": 206,
476
+ "se</w>": 741,
477
+ "sh": 795,
478
+ "so</w>": 630,
479
+ "sp": 992,
480
+ "ss": 673,
481
+ "st": 519,
482
+ "st</w>": 528,
483
+ "t": 85,
484
+ "t</w>": 208,
485
+ "te</w>": 954,
486
+ "ted</w>": 489,
487
+ "ter": 535,
488
+ "ter</w>": 505,
489
+ "th": 449,
490
+ "th</w>": 488,
491
+ "ther</w>": 576,
492
+ "ting</w>": 676,
493
+ "tion</w>": 570,
494
+ "tr": 619,
495
+ "ts</w>": 436,
496
+ "tt": 720,
497
+ "tur": 953,
498
+ "ty</w>": 821,
499
+ "u": 86,
500
+ "u</w>": 229,
501
+ "ub": 591,
502
+ "ubl": 842,
503
+ "uc": 490,
504
+ "ud": 538,
505
+ "ue</w>": 652,
506
+ "ug": 560,
507
+ "uil": 930,
508
+ "ul": 494,
509
+ "um": 532,
510
+ "um</w>": 644,
511
+ "un": 448,
512
+ "und</w>": 828,
513
+ "up": 833,
514
+ "up</w>": 700,
515
+ "ur": 413,
516
+ "ure</w>": 635,
517
+ "uring</w>": 702,
518
+ "ury</w>": 957,
519
+ "us": 438,
520
+ "us</w>": 622,
521
+ "ust</w>": 846,
522
+ "ut": 529,
523
+ "ut</w>": 527,
524
+ "v": 87,
525
+ "v</w>": 232,
526
+ "ve</w>": 567,
527
+ "vi": 866,
528
+ "w": 88,
529
+ "w</w>": 250,
530
+ "way</w>": 970,
531
+ "wn</w>": 999,
532
+ "x": 89,
533
+ "x</w>": 269,
534
+ "y": 90,
535
+ "y</w>": 211,
536
+ "yp": 993,
537
+ "z": 91,
538
+ "z</w>": 228,
539
+ "|": 92,
540
+ "|</w>": 304,
541
+ "}": 93,
542
+ "}</w>": 336,
543
+ "~": 94,
544
+ "~</w>": 343,
545
+ "¡": 95,
546
+ "¡</w>": 220,
547
+ "¢": 96,
548
+ "¢</w>": 306,
549
+ "£": 97,
550
+ "£</w>": 323,
551
+ "¤": 98,
552
+ "¤</w>": 292,
553
+ "¥": 99,
554
+ "¥</w>": 339,
555
+ "¦": 100,
556
+ "¦</w>": 303,
557
+ "§": 101,
558
+ "§</w>": 275,
559
+ "¨": 102,
560
+ "¨</w>": 282,
561
+ "©": 103,
562
+ "©</w>": 259,
563
+ "ª": 104,
564
+ "ª</w>": 286,
565
+ "«": 105,
566
+ "«</w>": 266,
567
+ "¬": 106,
568
+ "¬</w>": 319,
569
+ "®": 107,
570
+ "®</w>": 329,
571
+ "¯": 108,
572
+ "¯</w>": 287,
573
+ "°": 109,
574
+ "°</w>": 298,
575
+ "±": 110,
576
+ "±</w>": 200,
577
+ "²": 111,
578
+ "²</w>": 284,
579
+ "³": 112,
580
+ "³</w>": 272,
581
+ "´": 113,
582
+ "´</w>": 307,
583
+ "µ": 114,
584
+ "µ</w>": 261,
585
+ "¶": 115,
586
+ "¶</w>": 301,
587
+ "·": 116,
588
+ "·</w>": 326,
589
+ "¸": 117,
590
+ "¸</w>": 257,
591
+ "¹": 118,
592
+ "¹</w>": 241,
593
+ "º": 119,
594
+ "º</w>": 260,
595
+ "»": 120,
596
+ "»</w>": 247,
597
+ "¼": 121,
598
+ "¼</w>": 305,
599
+ "½": 122,
600
+ "½</w>": 294,
601
+ "¾": 123,
602
+ "¾</w>": 316,
603
+ "¿": 124,
604
+ "¿</w>": 271,
605
+ "Â": 125,
606
+ "Ã": 126,
607
+ "Ä": 127,
608
+ "Å": 128,
609
+ "Æ": 129,
610
+ "Ç": 130,
611
+ "È": 131,
612
+ "É": 132,
613
+ "Ê": 133,
614
+ "Ë": 134,
615
+ "Ì": 135,
616
+ "Í": 136,
617
+ "Î": 137,
618
+ "Ï": 138,
619
+ "Ð": 139,
620
+ "Ñ": 140,
621
+ "Ö": 141,
622
+ "×": 142,
623
+ "Ø": 143,
624
+ "Ù": 144,
625
+ "Ü": 145,
626
+ "à": 146,
627
+ "á": 147,
628
+ "â": 148,
629
+ "ã": 149,
630
+ "ä": 150,
631
+ "å": 151,
632
+ "æ": 152,
633
+ "ç": 153,
634
+ "è": 154,
635
+ "é": 155,
636
+ "ë": 156,
637
+ "ì": 157,
638
+ "ï": 158,
639
+ "Ċ": 159,
640
+ "Ċ</w>": 349,
641
+ "Ġ": 160,
642
+ "Ġ\"</w>": 401,
643
+ "Ġ'</w>": 431,
644
+ "Ġ(</w>": 475,
645
+ "Ġ)</w>": 474,
646
+ "Ġ,</w>": 360,
647
+ "Ġ.</w>": 365,
648
+ "Ġ0": 847,
649
+ "Ġ1": 411,
650
+ "Ġ18": 769,
651
+ "Ġ19": 492,
652
+ "Ġ199": 893,
653
+ "Ġ1</w>": 778,
654
+ "Ġ2": 462,
655
+ "Ġ20": 522,
656
+ "Ġ200": 620,
657
+ "Ġ201": 734,
658
+ "Ġ2</w>": 813,
659
+ "Ġ3": 735,
660
+ "Ġ3</w>": 888,
661
+ "Ġ4": 870,
662
+ "Ġ5": 907,
663
+ "Ġ5</w>": 990,
664
+ "Ġ:</w>": 637,
665
+ "Ġ;</w>": 615,
666
+ "Ġ</w>": 333,
667
+ "Ġ=</w>": 399,
668
+ "Ġ@": 417,
669
+ "Ġ@,@</w>": 755,
670
+ "Ġ@-@</w>": 440,
671
+ "Ġ@.@</w>": 696,
672
+ "ĠA": 409,
673
+ "ĠA</w>": 807,
674
+ "ĠAl": 716,
675
+ "ĠAm": 829,
676
+ "ĠAmeric": 958,
677
+ "ĠAn": 784,
678
+ "ĠAr": 894,
679
+ "ĠB": 432,
680
+ "ĠC": 410,
681
+ "ĠCh": 581,
682
+ "ĠCom": 904,
683
+ "ĠD": 464,
684
+ "ĠE": 500,
685
+ "ĠEn": 878,
686
+ "ĠF": 470,
687
+ "ĠG": 482,
688
+ "ĠH": 445,
689
+ "ĠHe</w>": 742,
690
+ "ĠI": 442,
691
+ "ĠI</w>": 827,
692
+ "ĠIn": 704,
693
+ "ĠIn</w>": 574,
694
+ "ĠIt</w>": 774,
695
+ "ĠJ": 491,
696
+ "ĠK": 548,
697
+ "ĠL": 484,
698
+ "ĠM": 423,
699
+ "ĠMar": 776,
700
+ "ĠN": 483,
701
+ "ĠO": 504,
702
+ "ĠP": 450,
703
+ "ĠPar": 967,
704
+ "ĠR": 459,
705
+ "ĠS": 403,
706
+ "ĠSh": 750,
707
+ "ĠSt": 590,
708
+ "ĠT": 396,
709
+ "ĠTh": 414,
710
+ "ĠThe</w>": 437,
711
+ "ĠThis</w>": 997,
712
+ "ĠU": 585,
713
+ "ĠUn": 773,
714
+ "ĠV": 617,
715
+ "ĠW": 479,
716
+ "ĠWh": 853,
717
+ "ĠY": 757,
718
+ "Ġa": 356,
719
+ "Ġa</w>": 394,
720
+ "Ġab": 653,
721
+ "Ġabout</w>": 899,
722
+ "Ġac": 583,
723
+ "Ġacc": 874,
724
+ "Ġad": 656,
725
+ "Ġafter</w>": 763,
726
+ "Ġag": 725,
727
+ "Ġal": 476,
728
+ "Ġalb": 991,
729
+ "Ġall</w>": 839,
730
+ "Ġalso</w>": 641,
731
+ "Ġan": 602,
732
+ "Ġan</w>": 562,
733
+ "Ġand</w>": 381,
734
+ "Ġapp": 711,
735
+ "Ġar": 507,
736
+ "Ġare</w>": 601,
737
+ "Ġas</w>": 454,
738
+ "Ġass": 947,
739
+ "Ġat</w>": 514,
740
+ "Ġatt": 788,
741
+ "Ġb": 371,
742
+ "Ġbe": 499,
743
+ "Ġbe</w>": 595,
744
+ "Ġbec": 706,
745
+ "Ġbeen</w>": 686,
746
+ "Ġbeg": 915,
747
+ "Ġbetween</w>": 914,
748
+ "Ġbo": 819,
749
+ "Ġbut</w>": 623,
750
+ "Ġby</w>": 473,
751
+ "Ġc": 368,
752
+ "Ġcent": 823,
753
+ "Ġch": 526,
754
+ "Ġchar": 822,
755
+ "Ġcl": 689,
756
+ "Ġcom": 509,
757
+ "Ġcomm": 707,
758
+ "Ġcomp": 616,
759
+ "Ġcon": 477,
760
+ "Ġcons": 841,
761
+ "Ġcont": 655,
762
+ "Ġcre": 931,
763
+ "Ġd": 387,
764
+ "Ġde": 627,
765
+ "Ġdec": 873,
766
+ "Ġdef": 965,
767
+ "Ġdes": 738,
768
+ "Ġdi": 892,
769
+ "Ġdis": 708,
770
+ "Ġduring</w>": 864,
771
+ "Ġe": 421,
772
+ "Ġear": 854,
773
+ "Ġel": 869,
774
+ "Ġen": 662,
775
+ "Ġev": 682,
776
+ "Ġex": 539,
777
+ "Ġexp": 867,
778
+ "Ġf": 370,
779
+ "Ġfe": 845,
780
+ "Ġfil": 913,
781
+ "Ġfin": 786,
782
+ "Ġfir": 599,
783
+ "Ġfirst</w>": 626,
784
+ "Ġfl": 877,
785
+ "Ġfor": 614,
786
+ "Ġfor</w>": 458,
787
+ "Ġform": 963,
788
+ "Ġfrom</w>": 503,
789
+ "Ġg": 430,
790
+ "Ġgame</w>": 895,
791
+ "Ġgen": 985,
792
+ "Ġgro": 890,
793
+ "Ġh": 380,
794
+ "Ġha": 485,
795
+ "Ġhad</w>": 566,
796
+ "Ġhas</w>": 679,
797
+ "Ġhave</w>": 667,
798
+ "Ġhe</w>": 558,
799
+ "Ġher</w>": 660,
800
+ "Ġhim</w>": 896,
801
+ "Ġhis</w>": 512,
802
+ "Ġi": 366,
803
+ "Ġimp": 909,
804
+ "Ġin": 429,
805
+ "Ġin</w>": 389,
806
+ "Ġinc": 618,
807
+ "Ġinclud": 761,
808
+ "Ġind": 983,
809
+ "Ġint": 628,
810
+ "Ġinter": 832,
811
+ "Ġinto</w>": 785,
812
+ "Ġis</w>": 480,
813
+ "Ġit</w>": 533,
814
+ "Ġits</w>": 642,
815
+ "Ġj": 723,
816
+ "Ġk": 564,
817
+ "Ġkn": 857,
818
+ "Ġl": 398,
819
+ "Ġlar": 962,
820
+ "Ġlater</w>": 936,
821
+ "Ġlea": 868,
822
+ "Ġm": 386,
823
+ "Ġmade</w>": 949,
824
+ "Ġme": 968,
825
+ "Ġmore</w>": 802,
826
+ "Ġmost</w>": 910,
827
+ "Ġmov": 956,
828
+ "Ġmus": 966,
829
+ "Ġn": 415,
830
+ "Ġnew</w>": 989,
831
+ "Ġno": 547,
832
+ "Ġnor": 978,
833
+ "Ġnot</w>": 632,
834
+ "Ġnum": 926,
835
+ "Ġo": 359,
836
+ "Ġof</w>": 373,
837
+ "Ġoff": 875,
838
+ "Ġon": 551,
839
+ "Ġon</w>": 456,
840
+ "Ġone</w>": 677,
841
+ "Ġonly</w>": 871,
842
+ "Ġor": 699,
843
+ "Ġor</w>": 645,
844
+ "Ġother</w>": 787,
845
+ "Ġout</w>": 925,
846
+ "Ġov": 729,
847
+ "Ġover</w>": 856,
848
+ "Ġp": 379,
849
+ "Ġpar": 636,
850
+ "Ġper": 736,
851
+ "Ġpl": 698,
852
+ "Ġpla": 697,
853
+ "Ġplay": 808,
854
+ "Ġpos": 859,
855
+ "Ġpr": 889,
856
+ "Ġpre": 749,
857
+ "Ġpres": 912,
858
+ "Ġpro": 506,
859
+ "Ġproduc": 934,
860
+ "Ġqu": 955,
861
+ "Ġr": 521,
862
+ "Ġra": 863,
863
+ "Ġre": 400,
864
+ "Ġrec": 597,
865
+ "Ġrecor": 919,
866
+ "Ġreg": 943,
867
+ "Ġrel": 900,
868
+ "Ġrele": 946,
869
+ "Ġrem": 848,
870
+ "Ġrep": 762,
871
+ "Ġres": 650,
872
+ "Ġro": 629,
873
+ "Ġs": 361,
874
+ "Ġsa": 905,
875
+ "Ġsc": 732,
876
+ "Ġse": 569,
877
+ "Ġseason</w>": 948,
878
+ "Ġsec": 781,
879
+ "Ġser": 740,
880
+ "Ġsev": 884,
881
+ "Ġsh": 513,
882
+ "Ġshe</w>": 862,
883
+ "Ġsp": 578,
884
+ "Ġspec": 940,
885
+ "Ġst": 446,
886
+ "Ġstar": 939,
887
+ "Ġsu": 770,
888
+ "Ġsub": 969,
889
+ "Ġsuc": 764,
890
+ "Ġsuch</w>": 981,
891
+ "Ġt": 354,
892
+ "Ġth": 355,
893
+ "Ġthan</w>": 918,
894
+ "Ġthat</w>": 452,
895
+ "Ġthe": 502,
896
+ "Ġthe</w>": 357,
897
+ "Ġtheir</w>": 621,
898
+ "Ġthem</w>": 998,
899
+ "Ġthey</w>": 727,
900
+ "Ġthis</w>": 705,
901
+ "Ġthree</w>": 902,
902
+ "Ġthroug": 923,
903
+ "Ġtime</w>": 783,
904
+ "Ġto</w>": 391,
905
+ "Ġtra": 836,
906
+ "Ġtw": 639,
907
+ "Ġtwo</w>": 688,
908
+ "Ġun": 544,
909
+ "Ġup</w>": 898,
910
+ "Ġus": 668,
911
+ "Ġused</w>": 988,
912
+ "Ġv": 495,
913
+ "Ġw": 363,
914
+ "Ġwas</w>": 422,
915
+ "Ġwere</w>": 525,
916
+ "Ġwh": 443,
917
+ "Ġwhen</w>": 851,
918
+ "Ġwhere</w>": 995,
919
+ "Ġwhich</w>": 573,
920
+ "Ġwhile</w>": 935,
921
+ "Ġwho</w>": 724,
922
+ "Ġwit": 451,
923
+ "Ġwith": 994,
924
+ "Ġwith</w>": 466,
925
+ "Ġwor": 643,
926
+ "Ġwould</w>": 801,
927
+ "Ġwrit": 937,
928
+ "Ġy": 580,
929
+ "Ġyear</w>": 987,
930
+ "Ġâ": 556,
931
+ "ĠâĢ": 565,
932
+ "ĠâĢĵ</w>": 631,
933
+ "ĠĊ</w>": 412,
934
+ "Ģ": 161,
935
+ "Ģ</w>": 223,
936
+ "ģ": 162,
937
+ "ģ</w>": 273,
938
+ "Ĥ": 163,
939
+ "Ĥ</w>": 262,
940
+ "ĥ": 164,
941
+ "ĥ</w>": 337,
942
+ "Ħ": 165,
943
+ "Ħ</w>": 278,
944
+ "ħ": 166,
945
+ "ħ</w>": 281,
946
+ "Ĩ": 167,
947
+ "Ĩ</w>": 308,
948
+ "ĩ": 168,
949
+ "ĩ</w>": 225,
950
+ "Ī": 169,
951
+ "Ī</w>": 221,
952
+ "ī": 170,
953
+ "ī</w>": 244,
954
+ "Ĭ": 171,
955
+ "Ĭ</w>": 315,
956
+ "ĭ": 172,
957
+ "ĭ</w>": 321,
958
+ "Į": 173,
959
+ "Į</w>": 324,
960
+ "į": 174,
961
+ "į</w>": 302,
962
+ "İ": 175,
963
+ "İ</w>": 249,
964
+ "ı": 176,
965
+ "ı</w>": 332,
966
+ "IJ": 177,
967
+ "IJ</w>": 295,
968
+ "ij": 178,
969
+ "ij</w>": 313,
970
+ "Ĵ": 179,
971
+ "Ĵ</w>": 328,
972
+ "ĵ": 180,
973
+ "ĵ</w>": 312,
974
+ "Ķ": 181,
975
+ "Ķ</w>": 256,
976
+ "ķ": 182,
977
+ "ķ</w>": 314,
978
+ "ĸ": 183,
979
+ "ĸ</w>": 277,
980
+ "Ĺ": 184,
981
+ "Ĺ</w>": 322,
982
+ "ĺ": 185,
983
+ "ĺ</w>": 285,
984
+ "Ļ": 186,
985
+ "Ļ</w>": 267,
986
+ "ļ": 187,
987
+ "ļ</w>": 290,
988
+ "Ľ": 188,
989
+ "Ľ</w>": 311,
990
+ "ľ": 189,
991
+ "ľ</w>": 299,
992
+ "Ŀ": 190,
993
+ "Ŀ</w>": 291,
994
+ "ŀ": 191,
995
+ "ŀ</w>": 293,
996
+ "Ł": 192,
997
+ "Ł</w>": 335,
998
+ "ł": 193,
999
+ "ł</w>": 252,
1000
+ "Ń": 194,
1001
+ "Ń</w>": 297
1002
+ }
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_decoder/config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_commit_hash": null,
4
+ "_diffusers_version": "0.29.1",
5
+ "_name_or_path": "/home/ubuntu/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-torch/snapshots/a88cdfbd91f96ec7f61eb7484b652ff0f4ee701d/vae",
6
+ "_use_default_values": [
7
+ "scaling_factor",
8
+ "use_post_quant_conv",
9
+ "force_upcast",
10
+ "latents_mean",
11
+ "latents_std",
12
+ "shift_factor",
13
+ "use_quant_conv"
14
+ ],
15
+ "act_fn": "silu",
16
+ "block_out_channels": [
17
+ 32,
18
+ 64
19
+ ],
20
+ "down_block_types": [
21
+ "DownEncoderBlock2D",
22
+ "DownEncoderBlock2D"
23
+ ],
24
+ "force_upcast": true,
25
+ "in_channels": 3,
26
+ "latent_channels": 4,
27
+ "latents_mean": null,
28
+ "latents_std": null,
29
+ "layers_per_block": 1,
30
+ "neuron": {
31
+ "auto_cast": "matmul",
32
+ "auto_cast_type": "bf16",
33
+ "compiler_type": "neuronx-cc",
34
+ "compiler_version": "2.13.66.0+6dfecc895",
35
+ "dynamic_batch_size": false,
36
+ "inline_weights_to_neff": true,
37
+ "input_names": [
38
+ "latent_sample"
39
+ ],
40
+ "model_type": "vae-decoder",
41
+ "optlevel": "2",
42
+ "output_attentions": false,
43
+ "output_hidden_states": false,
44
+ "output_names": [
45
+ "sample"
46
+ ],
47
+ "static_batch_size": 1,
48
+ "static_height": 32,
49
+ "static_num_channels": 4,
50
+ "static_width": 32
51
+ },
52
+ "norm_num_groups": 32,
53
+ "out_channels": 3,
54
+ "sample_size": 128,
55
+ "scaling_factor": 0.18215,
56
+ "shift_factor": null,
57
+ "task": "semantic-segmentation",
58
+ "transformers_version": null,
59
+ "up_block_types": [
60
+ "UpDecoderBlock2D",
61
+ "UpDecoderBlock2D"
62
+ ],
63
+ "use_post_quant_conv": true,
64
+ "use_quant_conv": true
65
+ }
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_decoder/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c1bf7c1ccf057ce99f128847670b06802981ad985687687d3118db83e7c46bf
3
+ size 1788977
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_encoder/config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_commit_hash": null,
4
+ "_diffusers_version": "0.29.1",
5
+ "_name_or_path": "/home/ubuntu/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-torch/snapshots/a88cdfbd91f96ec7f61eb7484b652ff0f4ee701d/vae",
6
+ "_use_default_values": [
7
+ "scaling_factor",
8
+ "use_post_quant_conv",
9
+ "force_upcast",
10
+ "latents_mean",
11
+ "latents_std",
12
+ "shift_factor",
13
+ "use_quant_conv"
14
+ ],
15
+ "act_fn": "silu",
16
+ "block_out_channels": [
17
+ 32,
18
+ 64
19
+ ],
20
+ "down_block_types": [
21
+ "DownEncoderBlock2D",
22
+ "DownEncoderBlock2D"
23
+ ],
24
+ "force_upcast": true,
25
+ "in_channels": 3,
26
+ "latent_channels": 4,
27
+ "latents_mean": null,
28
+ "latents_std": null,
29
+ "layers_per_block": 1,
30
+ "neuron": {
31
+ "auto_cast": "matmul",
32
+ "auto_cast_type": "bf16",
33
+ "compiler_type": "neuronx-cc",
34
+ "compiler_version": "2.13.66.0+6dfecc895",
35
+ "dynamic_batch_size": false,
36
+ "inline_weights_to_neff": true,
37
+ "input_names": [
38
+ "sample"
39
+ ],
40
+ "model_type": "vae-encoder",
41
+ "optlevel": "2",
42
+ "output_attentions": false,
43
+ "output_hidden_states": false,
44
+ "output_names": [
45
+ "latent_sample"
46
+ ],
47
+ "static_batch_size": 1,
48
+ "static_height": 64,
49
+ "static_num_channels": 3,
50
+ "static_width": 64
51
+ },
52
+ "norm_num_groups": 32,
53
+ "out_channels": 3,
54
+ "sample_size": 128,
55
+ "scaling_factor": 0.18215,
56
+ "shift_factor": null,
57
+ "task": "semantic-segmentation",
58
+ "transformers_version": null,
59
+ "up_block_types": [
60
+ "UpDecoderBlock2D",
61
+ "UpDecoderBlock2D"
62
+ ],
63
+ "use_post_quant_conv": true,
64
+ "use_quant_conv": true
65
+ }
neuronxcc-2.13.66.0+6dfecc895/MODULE_0058934baf38ad450d00/vae_encoder/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45cee6fe35941d6246c9042e09f3e151e6b918c6df8adebe6f4b29d0cfa664d1
3
+ size 1522733