sequelbox commited on
Commit
fbb33e5
·
1 Parent(s): 29a2213

Upload folder using huggingface_hub (#1)

Browse files

- 94af3249ee7d2f6b8be2c389a2a532e25abd3e942df5aa49194f06624d6ec7df (df15c8e7968db4a15de24735690f4a788b59ee08)
- 62ec0adb2e6eb61b84bf5753b8d55937675316bafc0e2fdf25d11d6795dfa25e (82f99016ca11306c72cf7870c5b47973a7025f84)
- 412d6cdaa4d4d74e8381bb114a41f7c6801177a2eb8de0cedd99f6dd1a6dbb6d (6c7dab0ec7726f28301dc8bbfa560ff13fdf7004)
- 848585d11936c4ae112bf93bdb59e5c759aa2a948a29f5b348a21692483ded0d (31c80b91151cec920f1e1e84442511cc0471a8be)
- b6261dfaf7c3a8aeabf16b4204d3fe0ab6a2973320e9698aa480f4c4fefc0bc8 (dc4a1e2d9e3026ae69a7331847617986f2b06484)
- b1190a762b308fca274cd5b94f07c056b291b57eb45e64c23c9f3eb0cb2eccb2 (e0cd4ed8bb9b1de55367bacd186b440055e3bc9c)
- 5c13dc3596290b66b847eb68af8eee320d5cdbedecd932ced9db3676f12116dd (7f5e4c9e784b130b829c0c3b06266221e6b80172)
- 63dc6b21fa0cb02b631da79d5c11168ccc310d4a7d03a71af390090340092d3c (4b295ca4ce303037c72b736676ac2627ebc91595)
- ebfd399034a4ddc90b959eb358013d1247f31ec237ee8b63fdf29bfaa0d440b5 (01ad1d15d43903ff0abd112ae43cec38e397663a)
- eb0a6533b36e84c8b2e7a36ebf5e5923b543c2a7653d628976333cbdc9e80aa0 (a6a3a3553a3406433b1adaa92c696fc41d3bec15)
- c20665c5f23317785cec4c9d0f7da76320d5236daba6c1a221334726d790eadd (1794d11175fcd29d7d50f1a897b64cfd9fab0137)
- ce858e8dc9990be378e9eb741afd4893a9a815a0e7f001cf9fbbc7a7a78984e9 (d1c1bd8260cb043c6c7ac7c6a295f49f3a42d7f9)
- ba3c47abdaaf7b18127dbedabac08e41a5615f84313f6e51150a6eb38ab4ec92 (1d4444e8c62513a7db86aa98c58d43dd86f2c250)
- 2a1440f1980ef647bc431f7d01a7fecd4311d733f161bad6670c4c06b3f93b02 (fbf5e925072bb59236fad297b687cfaad63bebec)
- 121a0045e764e701c2a4bb1b67be01fe422c5935e71f312e7cffdddf8c3ce91e (caebecd7f93cd533d7f519031cd0214d93e5d428)
- 00684905142fd5ca6093cae55cde0a0bf816a296294347cc63d59edfb648aed1 (f7e9ab61d55afdadd4ad9138a9f0f50da21c94bc)
- f2fca6496e27649572bb3d50976660aa6e1212bb29517e8bae7ea05766b57ad8 (40c9eeb13d757524bc0ba8e264ec44009e3a6ea0)
- 67271d62879cecdfd5ebe7180421781a589d111a944ae6915c018e1f2400884d (7c7cd95f70aae7657a1b8de6fc39ac09b1cce869)
- 3997b6ff6c69cb8c4fb2e8595c5d9aeb4cb813a9c26c482ad479d016f7d7e89b (ce4de135747ddf5e32415b6dc224d765b4e5ee2e)
- 09e08699082b878ed8d9a54acf2762db31c24d7edfa266a20640c042056ce41c (def6b545391b37f6e7d1eca6fa9a4d09f6114805)
- 9c2d4c0e364dbca0fff46e6fcc881682d7de168f3918563849a7fcf712d63ec0 (0a18b9779a44f50c0f9ed00cc4cb7e9dd5fd39de)
- 901c3c78b7f74b4022cd2745c381b08924dee6d5bb487b4eca4fe6ba382e8208 (7acd1f7688a10057cc3e2dcb68e710598a51d073)
- 8004e80664ae55918df6a582838a375c00a08fcd2d02a1d126f39c3a0742a36d (67aa82f8bfc304e58a2c17cdcf94970662d9567f)
- f8949fda5a613104702af6da5971c96c7fb3edd8c60d3eace9a6d55643480cce (3a03090755c5346ee85df97364d3a1bd736083b7)
- ebb0eb0f116d8704168107e0dc6b4448e1345782dee8d61d0a64cc46b0799466 (9fbe051ffbdc5e7bbd215575df943022018de970)
- 8785f45fc040fe79bd5cf21743c792e45d78f3197270e06da2d8c45e20533413 (698c045206713048d62b2c4d718b6a6ba99756ef)
- 741bdd74d3f2a553d717a94b1f0cd884ea959819cc455a33747ec14224ceace8 (1e9ced62984a72df54e2ad4f713e0a18c8b5ed68)
- ae44b464198fb42818814052f3f3b3881a7b33bbc5ebe633c1a71df83401fe38 (75d8e2058fcdd70fa913b1b28086e2fb9e8f86d4)
- 179c55eaa9f2e961752ab2fa32c63faa05f1e161396368566e663d641e9b2e95 (8345d3c8ab52c5fd8f25c4d1e9fd7f8baaddfc60)
- e2957fbd388fac7a32132f0e1b55975cbb769ba39f85a3e0db65af930604a609 (832219a75f6e30933a429bd59d0ddd7f2ffcd7d8)
- 6705812196fc694d4e1e9cf438d0e8e67ead54ecf0f0af5bdf5bf8e8f1e0eea1 (4d5629fe2ab00c24555feaa8793388d29463a265)
- 52837442612124c9c9bb5233b20375f107744a77620856b6fc0f1a60cc2b8c85 (7e871ffdb471d4cf98676d3ba288245692e12464)
- 13cafd840b0f1f35d94e8f743b29819b9b6763093851cd0a46a91aa1432e23be (a33e51a91c9f009a9d8fcb6f7bc107b202b67995)
- 3350024eeabd99f4dccc2c3824fc11da46ef3284199de7247fd1a079621696d7 (714c0a61e55148e9abeb0c26b00bb790e73ad78c)
- c9c9a18ead97df737c71124e5fcd02ac6e3360d0cf917c8a87e328f9c9fa1cc9 (e101b0649cd2ca454121c5c350aa85f04be16287)
- a103c940c961368b32ee2b306570bea3c6f02f223e3057d7a5e1ffb85412c700 (2b39a6facf0c93c37aa27b6a69ba31849b415f65)
- 372638ea9803605831171cac9add101fa21e78eb828b5d6c0cb5f794863e02e6 (b5436a2e078fe263ecef51ebc6dd8df2c21e557a)
- 039241a347695a6cbe01646105f544b8fafdf2c6b57a948bebe02b3b242e70c7 (b5ab169dbb361605c2956d14c90d38dfe16a24eb)
- d3464a3dda25bed02f33528d17eefa7a6174e2f310a42ff51cf081e59e18d00d (442f59e7e8208c650d6bcb2fb8a5e1b52d753f6b)
- 5860194d83aa6c31789555105659e6641529d8bf2ab998da9454a910a9ae2a2a (9f5ff064f5c0a48845b5b50b334202c81c1f39d1)
- fb8112f30fe67d8ed6127d87a546eaa5e2f4fcb41b1e60b0b615aaa13b75df20 (3f907711104336280aee10d3c92b400113b24115)
- 7399f9197a0ed39a9ae6fdffea87ef092d309c6f27dc096895929f20a32ab94f (c82f83089476a1ebef4bea363ed74562ae744140)
- ff1cf3a250ca5542cbeba0177fa3ff0c8e3cd8b106bce020d561bcc50c5255e7 (ebbb508ed49694f5a73451bd1c302e6005002777)
- b7340f108dfd263e1457bdf8cbf0747d97bf50e221a1b00ad9f6e066d1516ca3 (fee1e0f7582e1058d977d3c13f66a4ac5cfc5aaf)
- c64d4878e5b08398e3220973cfbcf653d51c7723c7945afaa5ac3189a7f866b3 (2335ab2c7a187d1bcc66df3ea354d098c0a2db10)
- fbe73ebc9846c7f284815109a4d5078243d7317feeaf9b18e4c1f892f565117f (0bd814d8a989082a4ecd8b451af509941d1c840b)
- cd5a0c152656f885ef2e0b70b9ef200ab1c1472f4d928e94b7a3098ac804b6bc (7b217ed8ea096c79bc41b61acb5938a259f38cad)
- a8cadbf0fe87dd61e102196e337da7b1a6e25ee009025ac6b159a583972cb55a (97c86e9dbc0be8598fbfa21a00e426529eb3df49)
- ac825f5998dec81cb71466d4bbc5c03cc83db88a27a9e870998928d45c245e58 (8cdcb62b0b3cfeb7abdd0d1e331c220deecd1dea)
- d5a8550526c35940bfcef1f0fb3af4065b96b847115e02c7d268ace2c98c0af3 (a73576e9b03235b4c4c657e6484eb5a68dfd7e93)
- fba5c2437f8af7d91bf3c30af0e9776747d0888ce69dcd3edd81c1097e7474e0 (7a2a960c2f1411cf506bf9181d1e900b48273e04)
- 90bb393b644f27c80abba53566ee0c8325ade8e86a08efc8b1f815a4fe6e8cdd (8948e27f1942cb9237360feb40f37061c2ad49a9)
- 3648ab95925cf7ea5ecba4525589ba464f4c321beca1f47b15154f075ba87fa1 (101e11c5a860fc7e8346c4da5f7bda4088ffbb72)
- fe904876f884f632ddcd034acec523bfb870b804775c3ed1a40389cd9dc67a7d (e8aae75634eb0b35aa811ecfad83cd123bec3c0e)
- 0024f1ce915d6e407473592476b05038f8cdb0709d2a4aa331c9f30a00d9e487 (92efb423c37759fd50aa0aea1db9e4c12f3aac98)
- 5998b27b46112b6ce6f2df515c9a1b9d212e3e75ffb7c35966680964b85a449d (a4a5d884f06a5cecee1395ff7a2d89d3283d646c)
- 61e15c2eff9b3564b8f076bcb650b9e9847d7aabc386eda7d77f10e8c49f1136 (856f0b2b9826d2b51bf171ea647d92e90fca715f)
- 3a2f406c7c2cc7de0cf4a198623824c979809a74ec4f26dad2aa47189b42215f (bb381012b149bd75e965712581ce299c0f544612)
- de3669e9b655c76dbc22feeee91d4caea1fbf028c9aacb565d435fdf2319c5d5 (8005c4cc90add8c3811683b2a55f0bb411165bc4)
- f0868b81529d08b281ef5b27f1d312282532a034d546729fd5d11129b47a1a51 (d4ba2af71396c2e88dcdcc7591e89b7b197d0d0d)
- 76d9822d3fe37ddc05aeede0658bd020de5c5917050cde9098ec4db175c12956 (1f64b191bbbca53912f938ed10b67e866c54a59d)

This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .ipynb_checkpoints/config-checkpoint.json +29 -0
  2. .ipynb_checkpoints/generation_config-checkpoint.json +8 -0
  3. .ipynb_checkpoints/model.safetensors.index-checkpoint.json +730 -0
  4. .ipynb_checkpoints/special_tokens_map-checkpoint.json +24 -0
  5. .ipynb_checkpoints/tokenizer-checkpoint.json +0 -0
  6. .ipynb_checkpoints/tokenizer_config-checkpoint.json +40 -0
  7. config.json +29 -0
  8. generation_config.json +8 -0
  9. model-00001-of-00061.safetensors +3 -0
  10. model-00002-of-00061.safetensors +3 -0
  11. model-00003-of-00061.safetensors +3 -0
  12. model-00004-of-00061.safetensors +3 -0
  13. model-00005-of-00061.safetensors +3 -0
  14. model-00006-of-00061.safetensors +3 -0
  15. model-00007-of-00061.safetensors +3 -0
  16. model-00008-of-00061.safetensors +3 -0
  17. model-00009-of-00061.safetensors +3 -0
  18. model-00010-of-00061.safetensors +3 -0
  19. model-00011-of-00061.safetensors +3 -0
  20. model-00012-of-00061.safetensors +3 -0
  21. model-00013-of-00061.safetensors +3 -0
  22. model-00014-of-00061.safetensors +3 -0
  23. model-00015-of-00061.safetensors +3 -0
  24. model-00016-of-00061.safetensors +3 -0
  25. model-00017-of-00061.safetensors +3 -0
  26. model-00018-of-00061.safetensors +3 -0
  27. model-00019-of-00061.safetensors +3 -0
  28. model-00020-of-00061.safetensors +3 -0
  29. model-00021-of-00061.safetensors +3 -0
  30. model-00022-of-00061.safetensors +3 -0
  31. model-00023-of-00061.safetensors +3 -0
  32. model-00024-of-00061.safetensors +3 -0
  33. model-00025-of-00061.safetensors +3 -0
  34. model-00026-of-00061.safetensors +3 -0
  35. model-00027-of-00061.safetensors +3 -0
  36. model-00028-of-00061.safetensors +3 -0
  37. model-00029-of-00061.safetensors +3 -0
  38. model-00030-of-00061.safetensors +3 -0
  39. model-00031-of-00061.safetensors +3 -0
  40. model-00032-of-00061.safetensors +3 -0
  41. model-00033-of-00061.safetensors +3 -0
  42. model-00034-of-00061.safetensors +3 -0
  43. model-00035-of-00061.safetensors +3 -0
  44. model-00036-of-00061.safetensors +3 -0
  45. model-00037-of-00061.safetensors +3 -0
  46. model-00038-of-00061.safetensors +3 -0
  47. model-00039-of-00061.safetensors +3 -0
  48. model-00040-of-00061.safetensors +3 -0
  49. model-00041-of-00061.safetensors +3 -0
  50. model-00042-of-00061.safetensors +3 -0
.ipynb_checkpoints/config-checkpoint.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "sequelbox/SpellBlade",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 8192,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 28672,
14
+ "max_position_embeddings": 4096,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 64,
17
+ "num_hidden_layers": 80,
18
+ "num_key_value_heads": 8,
19
+ "pad_token_id": 0,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_scaling": null,
23
+ "rope_theta": 10000.0,
24
+ "tie_word_embeddings": false,
25
+ "torch_dtype": "float32",
26
+ "transformers_version": "4.36.2",
27
+ "use_cache": false,
28
+ "vocab_size": 32000
29
+ }
.ipynb_checkpoints/generation_config-checkpoint.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.36.2",
7
+ "use_cache": false
8
+ }
.ipynb_checkpoints/model.safetensors.index-checkpoint.json ADDED
@@ -0,0 +1,730 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 275906592768
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00061-of-00061.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00061.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00061.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00061.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00061.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00061.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00061.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00061.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00061.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00061.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00061.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00002-of-00061.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00002-of-00061.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00061.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00002-of-00061.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00061.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00061.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00061.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00061.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00061.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00009-of-00061.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00009-of-00061.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00061.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00009-of-00061.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00061.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00061.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00061.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00061.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00061.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00010-of-00061.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00010-of-00061.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00061.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00009-of-00061.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00061.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00061.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00061.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00061.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00061.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00010-of-00061.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00010-of-00061.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00061.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00010-of-00061.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00061.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00061.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00061.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00061.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00061.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00011-of-00061.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00011-of-00061.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00061.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00011-of-00061.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00061.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00061.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00061.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00061.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00061.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00012-of-00061.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00012-of-00061.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00061.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00012-of-00061.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00061.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00061.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00061.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00061.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00061.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00013-of-00061.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00013-of-00061.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00061.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00012-of-00061.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00061.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00061.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00061.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00061.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00061.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00013-of-00061.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00013-of-00061.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00013-of-00061.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00013-of-00061.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00061.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00061.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00061.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00061.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00061.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00014-of-00061.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00014-of-00061.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00061.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00014-of-00061.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00061.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00061.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00061.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00061.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00061.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00015-of-00061.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00015-of-00061.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00061.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00015-of-00061.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00061.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00061.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00061.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00061.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00061.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00016-of-00061.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00016-of-00061.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00015-of-00061.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00015-of-00061.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00061.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00061.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00061.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00061.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00061.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00003-of-00061.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00003-of-00061.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00061.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00003-of-00061.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00061.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00061.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00061.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00061.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00061.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00016-of-00061.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00016-of-00061.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00016-of-00061.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00016-of-00061.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00061.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00061.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00061.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00061.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00061.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00017-of-00061.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00017-of-00061.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00061.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00017-of-00061.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00061.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00061.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00061.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00061.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00061.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00018-of-00061.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00018-of-00061.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00017-of-00061.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00018-of-00061.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00061.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00061.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00061.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00061.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00061.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00019-of-00061.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00019-of-00061.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00018-of-00061.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00018-of-00061.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00061.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00061.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00061.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00061.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00061.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00019-of-00061.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00019-of-00061.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00019-of-00061.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00019-of-00061.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00061.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00061.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00061.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00061.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00061.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00020-of-00061.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00020-of-00061.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00061.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00020-of-00061.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00061.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00061.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00061.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00061.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00061.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00021-of-00061.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00021-of-00061.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00061.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00021-of-00061.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00061.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00061.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00061.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00061.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00061.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00022-of-00061.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00022-of-00061.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00061.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00021-of-00061.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00061.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00061.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00061.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00061.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00061.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00022-of-00061.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00022-of-00061.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00061.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00022-of-00061.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00061.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00061.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00061.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00061.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00061.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00023-of-00061.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00023-of-00061.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00061.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00023-of-00061.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00061.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00061.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00061.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00061.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00061.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00004-of-00061.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00004-of-00061.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00061.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00003-of-00061.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00061.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00061.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00061.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00061.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00061.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00024-of-00061.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00024-of-00061.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00023-of-00061.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00024-of-00061.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00061.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00061.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00061.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00061.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00061.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00025-of-00061.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00025-of-00061.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00024-of-00061.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00024-of-00061.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00061.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00061.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00061.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00061.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00061.safetensors",
242
+ "model.layers.32.input_layernorm.weight": "model-00025-of-00061.safetensors",
243
+ "model.layers.32.mlp.down_proj.weight": "model-00025-of-00061.safetensors",
244
+ "model.layers.32.mlp.gate_proj.weight": "model-00025-of-00061.safetensors",
245
+ "model.layers.32.mlp.up_proj.weight": "model-00025-of-00061.safetensors",
246
+ "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00061.safetensors",
247
+ "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00061.safetensors",
248
+ "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00061.safetensors",
249
+ "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00061.safetensors",
250
+ "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00061.safetensors",
251
+ "model.layers.33.input_layernorm.weight": "model-00026-of-00061.safetensors",
252
+ "model.layers.33.mlp.down_proj.weight": "model-00026-of-00061.safetensors",
253
+ "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00061.safetensors",
254
+ "model.layers.33.mlp.up_proj.weight": "model-00026-of-00061.safetensors",
255
+ "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00061.safetensors",
256
+ "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00061.safetensors",
257
+ "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00061.safetensors",
258
+ "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00061.safetensors",
259
+ "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00061.safetensors",
260
+ "model.layers.34.input_layernorm.weight": "model-00027-of-00061.safetensors",
261
+ "model.layers.34.mlp.down_proj.weight": "model-00027-of-00061.safetensors",
262
+ "model.layers.34.mlp.gate_proj.weight": "model-00026-of-00061.safetensors",
263
+ "model.layers.34.mlp.up_proj.weight": "model-00027-of-00061.safetensors",
264
+ "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00061.safetensors",
265
+ "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00061.safetensors",
266
+ "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00061.safetensors",
267
+ "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00061.safetensors",
268
+ "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00061.safetensors",
269
+ "model.layers.35.input_layernorm.weight": "model-00028-of-00061.safetensors",
270
+ "model.layers.35.mlp.down_proj.weight": "model-00028-of-00061.safetensors",
271
+ "model.layers.35.mlp.gate_proj.weight": "model-00027-of-00061.safetensors",
272
+ "model.layers.35.mlp.up_proj.weight": "model-00027-of-00061.safetensors",
273
+ "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00061.safetensors",
274
+ "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00061.safetensors",
275
+ "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00061.safetensors",
276
+ "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00061.safetensors",
277
+ "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00061.safetensors",
278
+ "model.layers.36.input_layernorm.weight": "model-00028-of-00061.safetensors",
279
+ "model.layers.36.mlp.down_proj.weight": "model-00028-of-00061.safetensors",
280
+ "model.layers.36.mlp.gate_proj.weight": "model-00028-of-00061.safetensors",
281
+ "model.layers.36.mlp.up_proj.weight": "model-00028-of-00061.safetensors",
282
+ "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00061.safetensors",
283
+ "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00061.safetensors",
284
+ "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00061.safetensors",
285
+ "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00061.safetensors",
286
+ "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00061.safetensors",
287
+ "model.layers.37.input_layernorm.weight": "model-00029-of-00061.safetensors",
288
+ "model.layers.37.mlp.down_proj.weight": "model-00029-of-00061.safetensors",
289
+ "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00061.safetensors",
290
+ "model.layers.37.mlp.up_proj.weight": "model-00029-of-00061.safetensors",
291
+ "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00061.safetensors",
292
+ "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00061.safetensors",
293
+ "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00061.safetensors",
294
+ "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00061.safetensors",
295
+ "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00061.safetensors",
296
+ "model.layers.38.input_layernorm.weight": "model-00030-of-00061.safetensors",
297
+ "model.layers.38.mlp.down_proj.weight": "model-00030-of-00061.safetensors",
298
+ "model.layers.38.mlp.gate_proj.weight": "model-00029-of-00061.safetensors",
299
+ "model.layers.38.mlp.up_proj.weight": "model-00030-of-00061.safetensors",
300
+ "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00061.safetensors",
301
+ "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00061.safetensors",
302
+ "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00061.safetensors",
303
+ "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00061.safetensors",
304
+ "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00061.safetensors",
305
+ "model.layers.39.input_layernorm.weight": "model-00031-of-00061.safetensors",
306
+ "model.layers.39.mlp.down_proj.weight": "model-00031-of-00061.safetensors",
307
+ "model.layers.39.mlp.gate_proj.weight": "model-00030-of-00061.safetensors",
308
+ "model.layers.39.mlp.up_proj.weight": "model-00030-of-00061.safetensors",
309
+ "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00061.safetensors",
310
+ "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00061.safetensors",
311
+ "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00061.safetensors",
312
+ "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00061.safetensors",
313
+ "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00061.safetensors",
314
+ "model.layers.4.input_layernorm.weight": "model-00004-of-00061.safetensors",
315
+ "model.layers.4.mlp.down_proj.weight": "model-00004-of-00061.safetensors",
316
+ "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00061.safetensors",
317
+ "model.layers.4.mlp.up_proj.weight": "model-00004-of-00061.safetensors",
318
+ "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00061.safetensors",
319
+ "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00061.safetensors",
320
+ "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00061.safetensors",
321
+ "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00061.safetensors",
322
+ "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00061.safetensors",
323
+ "model.layers.40.input_layernorm.weight": "model-00031-of-00061.safetensors",
324
+ "model.layers.40.mlp.down_proj.weight": "model-00031-of-00061.safetensors",
325
+ "model.layers.40.mlp.gate_proj.weight": "model-00031-of-00061.safetensors",
326
+ "model.layers.40.mlp.up_proj.weight": "model-00031-of-00061.safetensors",
327
+ "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00061.safetensors",
328
+ "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00061.safetensors",
329
+ "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00061.safetensors",
330
+ "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00061.safetensors",
331
+ "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00061.safetensors",
332
+ "model.layers.41.input_layernorm.weight": "model-00032-of-00061.safetensors",
333
+ "model.layers.41.mlp.down_proj.weight": "model-00032-of-00061.safetensors",
334
+ "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00061.safetensors",
335
+ "model.layers.41.mlp.up_proj.weight": "model-00032-of-00061.safetensors",
336
+ "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00061.safetensors",
337
+ "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00061.safetensors",
338
+ "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00061.safetensors",
339
+ "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00061.safetensors",
340
+ "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00061.safetensors",
341
+ "model.layers.42.input_layernorm.weight": "model-00033-of-00061.safetensors",
342
+ "model.layers.42.mlp.down_proj.weight": "model-00033-of-00061.safetensors",
343
+ "model.layers.42.mlp.gate_proj.weight": "model-00032-of-00061.safetensors",
344
+ "model.layers.42.mlp.up_proj.weight": "model-00033-of-00061.safetensors",
345
+ "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00061.safetensors",
346
+ "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00061.safetensors",
347
+ "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00061.safetensors",
348
+ "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00061.safetensors",
349
+ "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00061.safetensors",
350
+ "model.layers.43.input_layernorm.weight": "model-00034-of-00061.safetensors",
351
+ "model.layers.43.mlp.down_proj.weight": "model-00034-of-00061.safetensors",
352
+ "model.layers.43.mlp.gate_proj.weight": "model-00033-of-00061.safetensors",
353
+ "model.layers.43.mlp.up_proj.weight": "model-00033-of-00061.safetensors",
354
+ "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00061.safetensors",
355
+ "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00061.safetensors",
356
+ "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00061.safetensors",
357
+ "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00061.safetensors",
358
+ "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00061.safetensors",
359
+ "model.layers.44.input_layernorm.weight": "model-00034-of-00061.safetensors",
360
+ "model.layers.44.mlp.down_proj.weight": "model-00034-of-00061.safetensors",
361
+ "model.layers.44.mlp.gate_proj.weight": "model-00034-of-00061.safetensors",
362
+ "model.layers.44.mlp.up_proj.weight": "model-00034-of-00061.safetensors",
363
+ "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00061.safetensors",
364
+ "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00061.safetensors",
365
+ "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00061.safetensors",
366
+ "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00061.safetensors",
367
+ "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00061.safetensors",
368
+ "model.layers.45.input_layernorm.weight": "model-00035-of-00061.safetensors",
369
+ "model.layers.45.mlp.down_proj.weight": "model-00035-of-00061.safetensors",
370
+ "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00061.safetensors",
371
+ "model.layers.45.mlp.up_proj.weight": "model-00035-of-00061.safetensors",
372
+ "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00061.safetensors",
373
+ "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00061.safetensors",
374
+ "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00061.safetensors",
375
+ "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00061.safetensors",
376
+ "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00061.safetensors",
377
+ "model.layers.46.input_layernorm.weight": "model-00036-of-00061.safetensors",
378
+ "model.layers.46.mlp.down_proj.weight": "model-00036-of-00061.safetensors",
379
+ "model.layers.46.mlp.gate_proj.weight": "model-00035-of-00061.safetensors",
380
+ "model.layers.46.mlp.up_proj.weight": "model-00036-of-00061.safetensors",
381
+ "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00061.safetensors",
382
+ "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00061.safetensors",
383
+ "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00061.safetensors",
384
+ "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00061.safetensors",
385
+ "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00061.safetensors",
386
+ "model.layers.47.input_layernorm.weight": "model-00037-of-00061.safetensors",
387
+ "model.layers.47.mlp.down_proj.weight": "model-00037-of-00061.safetensors",
388
+ "model.layers.47.mlp.gate_proj.weight": "model-00036-of-00061.safetensors",
389
+ "model.layers.47.mlp.up_proj.weight": "model-00036-of-00061.safetensors",
390
+ "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00061.safetensors",
391
+ "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00061.safetensors",
392
+ "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00061.safetensors",
393
+ "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00061.safetensors",
394
+ "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00061.safetensors",
395
+ "model.layers.48.input_layernorm.weight": "model-00037-of-00061.safetensors",
396
+ "model.layers.48.mlp.down_proj.weight": "model-00037-of-00061.safetensors",
397
+ "model.layers.48.mlp.gate_proj.weight": "model-00037-of-00061.safetensors",
398
+ "model.layers.48.mlp.up_proj.weight": "model-00037-of-00061.safetensors",
399
+ "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00061.safetensors",
400
+ "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00061.safetensors",
401
+ "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00061.safetensors",
402
+ "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00061.safetensors",
403
+ "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00061.safetensors",
404
+ "model.layers.49.input_layernorm.weight": "model-00038-of-00061.safetensors",
405
+ "model.layers.49.mlp.down_proj.weight": "model-00038-of-00061.safetensors",
406
+ "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00061.safetensors",
407
+ "model.layers.49.mlp.up_proj.weight": "model-00038-of-00061.safetensors",
408
+ "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00061.safetensors",
409
+ "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00061.safetensors",
410
+ "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00061.safetensors",
411
+ "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00061.safetensors",
412
+ "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00061.safetensors",
413
+ "model.layers.5.input_layernorm.weight": "model-00005-of-00061.safetensors",
414
+ "model.layers.5.mlp.down_proj.weight": "model-00005-of-00061.safetensors",
415
+ "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00061.safetensors",
416
+ "model.layers.5.mlp.up_proj.weight": "model-00005-of-00061.safetensors",
417
+ "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00061.safetensors",
418
+ "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00061.safetensors",
419
+ "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00061.safetensors",
420
+ "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00061.safetensors",
421
+ "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00061.safetensors",
422
+ "model.layers.50.input_layernorm.weight": "model-00039-of-00061.safetensors",
423
+ "model.layers.50.mlp.down_proj.weight": "model-00039-of-00061.safetensors",
424
+ "model.layers.50.mlp.gate_proj.weight": "model-00038-of-00061.safetensors",
425
+ "model.layers.50.mlp.up_proj.weight": "model-00039-of-00061.safetensors",
426
+ "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00061.safetensors",
427
+ "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00061.safetensors",
428
+ "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00061.safetensors",
429
+ "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00061.safetensors",
430
+ "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00061.safetensors",
431
+ "model.layers.51.input_layernorm.weight": "model-00040-of-00061.safetensors",
432
+ "model.layers.51.mlp.down_proj.weight": "model-00040-of-00061.safetensors",
433
+ "model.layers.51.mlp.gate_proj.weight": "model-00039-of-00061.safetensors",
434
+ "model.layers.51.mlp.up_proj.weight": "model-00039-of-00061.safetensors",
435
+ "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00061.safetensors",
436
+ "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00061.safetensors",
437
+ "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00061.safetensors",
438
+ "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00061.safetensors",
439
+ "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00061.safetensors",
440
+ "model.layers.52.input_layernorm.weight": "model-00040-of-00061.safetensors",
441
+ "model.layers.52.mlp.down_proj.weight": "model-00040-of-00061.safetensors",
442
+ "model.layers.52.mlp.gate_proj.weight": "model-00040-of-00061.safetensors",
443
+ "model.layers.52.mlp.up_proj.weight": "model-00040-of-00061.safetensors",
444
+ "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00061.safetensors",
445
+ "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00061.safetensors",
446
+ "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00061.safetensors",
447
+ "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00061.safetensors",
448
+ "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00061.safetensors",
449
+ "model.layers.53.input_layernorm.weight": "model-00041-of-00061.safetensors",
450
+ "model.layers.53.mlp.down_proj.weight": "model-00041-of-00061.safetensors",
451
+ "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00061.safetensors",
452
+ "model.layers.53.mlp.up_proj.weight": "model-00041-of-00061.safetensors",
453
+ "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00061.safetensors",
454
+ "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00061.safetensors",
455
+ "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00061.safetensors",
456
+ "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00061.safetensors",
457
+ "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00061.safetensors",
458
+ "model.layers.54.input_layernorm.weight": "model-00042-of-00061.safetensors",
459
+ "model.layers.54.mlp.down_proj.weight": "model-00042-of-00061.safetensors",
460
+ "model.layers.54.mlp.gate_proj.weight": "model-00041-of-00061.safetensors",
461
+ "model.layers.54.mlp.up_proj.weight": "model-00042-of-00061.safetensors",
462
+ "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00061.safetensors",
463
+ "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00061.safetensors",
464
+ "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00061.safetensors",
465
+ "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00061.safetensors",
466
+ "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00061.safetensors",
467
+ "model.layers.55.input_layernorm.weight": "model-00043-of-00061.safetensors",
468
+ "model.layers.55.mlp.down_proj.weight": "model-00043-of-00061.safetensors",
469
+ "model.layers.55.mlp.gate_proj.weight": "model-00042-of-00061.safetensors",
470
+ "model.layers.55.mlp.up_proj.weight": "model-00042-of-00061.safetensors",
471
+ "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00061.safetensors",
472
+ "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00061.safetensors",
473
+ "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00061.safetensors",
474
+ "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00061.safetensors",
475
+ "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00061.safetensors",
476
+ "model.layers.56.input_layernorm.weight": "model-00043-of-00061.safetensors",
477
+ "model.layers.56.mlp.down_proj.weight": "model-00043-of-00061.safetensors",
478
+ "model.layers.56.mlp.gate_proj.weight": "model-00043-of-00061.safetensors",
479
+ "model.layers.56.mlp.up_proj.weight": "model-00043-of-00061.safetensors",
480
+ "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00061.safetensors",
481
+ "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00061.safetensors",
482
+ "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00061.safetensors",
483
+ "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00061.safetensors",
484
+ "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00061.safetensors",
485
+ "model.layers.57.input_layernorm.weight": "model-00044-of-00061.safetensors",
486
+ "model.layers.57.mlp.down_proj.weight": "model-00044-of-00061.safetensors",
487
+ "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00061.safetensors",
488
+ "model.layers.57.mlp.up_proj.weight": "model-00044-of-00061.safetensors",
489
+ "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00061.safetensors",
490
+ "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00061.safetensors",
491
+ "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00061.safetensors",
492
+ "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00061.safetensors",
493
+ "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00061.safetensors",
494
+ "model.layers.58.input_layernorm.weight": "model-00045-of-00061.safetensors",
495
+ "model.layers.58.mlp.down_proj.weight": "model-00045-of-00061.safetensors",
496
+ "model.layers.58.mlp.gate_proj.weight": "model-00044-of-00061.safetensors",
497
+ "model.layers.58.mlp.up_proj.weight": "model-00045-of-00061.safetensors",
498
+ "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00061.safetensors",
499
+ "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00061.safetensors",
500
+ "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00061.safetensors",
501
+ "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00061.safetensors",
502
+ "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00061.safetensors",
503
+ "model.layers.59.input_layernorm.weight": "model-00046-of-00061.safetensors",
504
+ "model.layers.59.mlp.down_proj.weight": "model-00046-of-00061.safetensors",
505
+ "model.layers.59.mlp.gate_proj.weight": "model-00045-of-00061.safetensors",
506
+ "model.layers.59.mlp.up_proj.weight": "model-00045-of-00061.safetensors",
507
+ "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00061.safetensors",
508
+ "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00061.safetensors",
509
+ "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00061.safetensors",
510
+ "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00061.safetensors",
511
+ "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00061.safetensors",
512
+ "model.layers.6.input_layernorm.weight": "model-00006-of-00061.safetensors",
513
+ "model.layers.6.mlp.down_proj.weight": "model-00006-of-00061.safetensors",
514
+ "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00061.safetensors",
515
+ "model.layers.6.mlp.up_proj.weight": "model-00006-of-00061.safetensors",
516
+ "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00061.safetensors",
517
+ "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00061.safetensors",
518
+ "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00061.safetensors",
519
+ "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00061.safetensors",
520
+ "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00061.safetensors",
521
+ "model.layers.60.input_layernorm.weight": "model-00046-of-00061.safetensors",
522
+ "model.layers.60.mlp.down_proj.weight": "model-00046-of-00061.safetensors",
523
+ "model.layers.60.mlp.gate_proj.weight": "model-00046-of-00061.safetensors",
524
+ "model.layers.60.mlp.up_proj.weight": "model-00046-of-00061.safetensors",
525
+ "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00061.safetensors",
526
+ "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00061.safetensors",
527
+ "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00061.safetensors",
528
+ "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00061.safetensors",
529
+ "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00061.safetensors",
530
+ "model.layers.61.input_layernorm.weight": "model-00047-of-00061.safetensors",
531
+ "model.layers.61.mlp.down_proj.weight": "model-00047-of-00061.safetensors",
532
+ "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00061.safetensors",
533
+ "model.layers.61.mlp.up_proj.weight": "model-00047-of-00061.safetensors",
534
+ "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00061.safetensors",
535
+ "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00061.safetensors",
536
+ "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00061.safetensors",
537
+ "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00061.safetensors",
538
+ "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00061.safetensors",
539
+ "model.layers.62.input_layernorm.weight": "model-00048-of-00061.safetensors",
540
+ "model.layers.62.mlp.down_proj.weight": "model-00048-of-00061.safetensors",
541
+ "model.layers.62.mlp.gate_proj.weight": "model-00047-of-00061.safetensors",
542
+ "model.layers.62.mlp.up_proj.weight": "model-00048-of-00061.safetensors",
543
+ "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00061.safetensors",
544
+ "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00061.safetensors",
545
+ "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00061.safetensors",
546
+ "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00061.safetensors",
547
+ "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00061.safetensors",
548
+ "model.layers.63.input_layernorm.weight": "model-00049-of-00061.safetensors",
549
+ "model.layers.63.mlp.down_proj.weight": "model-00049-of-00061.safetensors",
550
+ "model.layers.63.mlp.gate_proj.weight": "model-00048-of-00061.safetensors",
551
+ "model.layers.63.mlp.up_proj.weight": "model-00048-of-00061.safetensors",
552
+ "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00061.safetensors",
553
+ "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00061.safetensors",
554
+ "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00061.safetensors",
555
+ "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00061.safetensors",
556
+ "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00061.safetensors",
557
+ "model.layers.64.input_layernorm.weight": "model-00049-of-00061.safetensors",
558
+ "model.layers.64.mlp.down_proj.weight": "model-00049-of-00061.safetensors",
559
+ "model.layers.64.mlp.gate_proj.weight": "model-00049-of-00061.safetensors",
560
+ "model.layers.64.mlp.up_proj.weight": "model-00049-of-00061.safetensors",
561
+ "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00061.safetensors",
562
+ "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00061.safetensors",
563
+ "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00061.safetensors",
564
+ "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00061.safetensors",
565
+ "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00061.safetensors",
566
+ "model.layers.65.input_layernorm.weight": "model-00050-of-00061.safetensors",
567
+ "model.layers.65.mlp.down_proj.weight": "model-00050-of-00061.safetensors",
568
+ "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00061.safetensors",
569
+ "model.layers.65.mlp.up_proj.weight": "model-00050-of-00061.safetensors",
570
+ "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00061.safetensors",
571
+ "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00061.safetensors",
572
+ "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00061.safetensors",
573
+ "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00061.safetensors",
574
+ "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00061.safetensors",
575
+ "model.layers.66.input_layernorm.weight": "model-00051-of-00061.safetensors",
576
+ "model.layers.66.mlp.down_proj.weight": "model-00051-of-00061.safetensors",
577
+ "model.layers.66.mlp.gate_proj.weight": "model-00050-of-00061.safetensors",
578
+ "model.layers.66.mlp.up_proj.weight": "model-00051-of-00061.safetensors",
579
+ "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00061.safetensors",
580
+ "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00061.safetensors",
581
+ "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00061.safetensors",
582
+ "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00061.safetensors",
583
+ "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00061.safetensors",
584
+ "model.layers.67.input_layernorm.weight": "model-00052-of-00061.safetensors",
585
+ "model.layers.67.mlp.down_proj.weight": "model-00052-of-00061.safetensors",
586
+ "model.layers.67.mlp.gate_proj.weight": "model-00051-of-00061.safetensors",
587
+ "model.layers.67.mlp.up_proj.weight": "model-00051-of-00061.safetensors",
588
+ "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00061.safetensors",
589
+ "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00061.safetensors",
590
+ "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00061.safetensors",
591
+ "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00061.safetensors",
592
+ "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00061.safetensors",
593
+ "model.layers.68.input_layernorm.weight": "model-00052-of-00061.safetensors",
594
+ "model.layers.68.mlp.down_proj.weight": "model-00052-of-00061.safetensors",
595
+ "model.layers.68.mlp.gate_proj.weight": "model-00052-of-00061.safetensors",
596
+ "model.layers.68.mlp.up_proj.weight": "model-00052-of-00061.safetensors",
597
+ "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00061.safetensors",
598
+ "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00061.safetensors",
599
+ "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00061.safetensors",
600
+ "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00061.safetensors",
601
+ "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00061.safetensors",
602
+ "model.layers.69.input_layernorm.weight": "model-00053-of-00061.safetensors",
603
+ "model.layers.69.mlp.down_proj.weight": "model-00053-of-00061.safetensors",
604
+ "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00061.safetensors",
605
+ "model.layers.69.mlp.up_proj.weight": "model-00053-of-00061.safetensors",
606
+ "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00061.safetensors",
607
+ "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00061.safetensors",
608
+ "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00061.safetensors",
609
+ "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00061.safetensors",
610
+ "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00061.safetensors",
611
+ "model.layers.7.input_layernorm.weight": "model-00007-of-00061.safetensors",
612
+ "model.layers.7.mlp.down_proj.weight": "model-00007-of-00061.safetensors",
613
+ "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00061.safetensors",
614
+ "model.layers.7.mlp.up_proj.weight": "model-00006-of-00061.safetensors",
615
+ "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00061.safetensors",
616
+ "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00061.safetensors",
617
+ "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00061.safetensors",
618
+ "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00061.safetensors",
619
+ "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00061.safetensors",
620
+ "model.layers.70.input_layernorm.weight": "model-00054-of-00061.safetensors",
621
+ "model.layers.70.mlp.down_proj.weight": "model-00054-of-00061.safetensors",
622
+ "model.layers.70.mlp.gate_proj.weight": "model-00053-of-00061.safetensors",
623
+ "model.layers.70.mlp.up_proj.weight": "model-00054-of-00061.safetensors",
624
+ "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00061.safetensors",
625
+ "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00061.safetensors",
626
+ "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00061.safetensors",
627
+ "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00061.safetensors",
628
+ "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00061.safetensors",
629
+ "model.layers.71.input_layernorm.weight": "model-00055-of-00061.safetensors",
630
+ "model.layers.71.mlp.down_proj.weight": "model-00055-of-00061.safetensors",
631
+ "model.layers.71.mlp.gate_proj.weight": "model-00054-of-00061.safetensors",
632
+ "model.layers.71.mlp.up_proj.weight": "model-00054-of-00061.safetensors",
633
+ "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00061.safetensors",
634
+ "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00061.safetensors",
635
+ "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00061.safetensors",
636
+ "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00061.safetensors",
637
+ "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00061.safetensors",
638
+ "model.layers.72.input_layernorm.weight": "model-00055-of-00061.safetensors",
639
+ "model.layers.72.mlp.down_proj.weight": "model-00055-of-00061.safetensors",
640
+ "model.layers.72.mlp.gate_proj.weight": "model-00055-of-00061.safetensors",
641
+ "model.layers.72.mlp.up_proj.weight": "model-00055-of-00061.safetensors",
642
+ "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00061.safetensors",
643
+ "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00061.safetensors",
644
+ "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00061.safetensors",
645
+ "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00061.safetensors",
646
+ "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00061.safetensors",
647
+ "model.layers.73.input_layernorm.weight": "model-00056-of-00061.safetensors",
648
+ "model.layers.73.mlp.down_proj.weight": "model-00056-of-00061.safetensors",
649
+ "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00061.safetensors",
650
+ "model.layers.73.mlp.up_proj.weight": "model-00056-of-00061.safetensors",
651
+ "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00061.safetensors",
652
+ "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00061.safetensors",
653
+ "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00061.safetensors",
654
+ "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00061.safetensors",
655
+ "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00061.safetensors",
656
+ "model.layers.74.input_layernorm.weight": "model-00057-of-00061.safetensors",
657
+ "model.layers.74.mlp.down_proj.weight": "model-00057-of-00061.safetensors",
658
+ "model.layers.74.mlp.gate_proj.weight": "model-00056-of-00061.safetensors",
659
+ "model.layers.74.mlp.up_proj.weight": "model-00057-of-00061.safetensors",
660
+ "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00061.safetensors",
661
+ "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00061.safetensors",
662
+ "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00061.safetensors",
663
+ "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00061.safetensors",
664
+ "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00061.safetensors",
665
+ "model.layers.75.input_layernorm.weight": "model-00058-of-00061.safetensors",
666
+ "model.layers.75.mlp.down_proj.weight": "model-00058-of-00061.safetensors",
667
+ "model.layers.75.mlp.gate_proj.weight": "model-00057-of-00061.safetensors",
668
+ "model.layers.75.mlp.up_proj.weight": "model-00057-of-00061.safetensors",
669
+ "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00061.safetensors",
670
+ "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00061.safetensors",
671
+ "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00061.safetensors",
672
+ "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00061.safetensors",
673
+ "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00061.safetensors",
674
+ "model.layers.76.input_layernorm.weight": "model-00058-of-00061.safetensors",
675
+ "model.layers.76.mlp.down_proj.weight": "model-00058-of-00061.safetensors",
676
+ "model.layers.76.mlp.gate_proj.weight": "model-00058-of-00061.safetensors",
677
+ "model.layers.76.mlp.up_proj.weight": "model-00058-of-00061.safetensors",
678
+ "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00061.safetensors",
679
+ "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00061.safetensors",
680
+ "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00061.safetensors",
681
+ "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00061.safetensors",
682
+ "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00061.safetensors",
683
+ "model.layers.77.input_layernorm.weight": "model-00059-of-00061.safetensors",
684
+ "model.layers.77.mlp.down_proj.weight": "model-00059-of-00061.safetensors",
685
+ "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00061.safetensors",
686
+ "model.layers.77.mlp.up_proj.weight": "model-00059-of-00061.safetensors",
687
+ "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00061.safetensors",
688
+ "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00061.safetensors",
689
+ "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00061.safetensors",
690
+ "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00061.safetensors",
691
+ "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00061.safetensors",
692
+ "model.layers.78.input_layernorm.weight": "model-00060-of-00061.safetensors",
693
+ "model.layers.78.mlp.down_proj.weight": "model-00060-of-00061.safetensors",
694
+ "model.layers.78.mlp.gate_proj.weight": "model-00059-of-00061.safetensors",
695
+ "model.layers.78.mlp.up_proj.weight": "model-00060-of-00061.safetensors",
696
+ "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00061.safetensors",
697
+ "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00061.safetensors",
698
+ "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00061.safetensors",
699
+ "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00061.safetensors",
700
+ "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00061.safetensors",
701
+ "model.layers.79.input_layernorm.weight": "model-00061-of-00061.safetensors",
702
+ "model.layers.79.mlp.down_proj.weight": "model-00061-of-00061.safetensors",
703
+ "model.layers.79.mlp.gate_proj.weight": "model-00060-of-00061.safetensors",
704
+ "model.layers.79.mlp.up_proj.weight": "model-00060-of-00061.safetensors",
705
+ "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00061.safetensors",
706
+ "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00061.safetensors",
707
+ "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00061.safetensors",
708
+ "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00061.safetensors",
709
+ "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00061.safetensors",
710
+ "model.layers.8.input_layernorm.weight": "model-00007-of-00061.safetensors",
711
+ "model.layers.8.mlp.down_proj.weight": "model-00007-of-00061.safetensors",
712
+ "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00061.safetensors",
713
+ "model.layers.8.mlp.up_proj.weight": "model-00007-of-00061.safetensors",
714
+ "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00061.safetensors",
715
+ "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00061.safetensors",
716
+ "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00061.safetensors",
717
+ "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00061.safetensors",
718
+ "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00061.safetensors",
719
+ "model.layers.9.input_layernorm.weight": "model-00008-of-00061.safetensors",
720
+ "model.layers.9.mlp.down_proj.weight": "model-00008-of-00061.safetensors",
721
+ "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00061.safetensors",
722
+ "model.layers.9.mlp.up_proj.weight": "model-00008-of-00061.safetensors",
723
+ "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00061.safetensors",
724
+ "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00061.safetensors",
725
+ "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00061.safetensors",
726
+ "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00061.safetensors",
727
+ "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00061.safetensors",
728
+ "model.norm.weight": "model-00061-of-00061.safetensors"
729
+ }
730
+ }
.ipynb_checkpoints/special_tokens_map-checkpoint.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
.ipynb_checkpoints/tokenizer-checkpoint.json ADDED
The diff for this file is too large to render. See raw diff
 
.ipynb_checkpoints/tokenizer_config-checkpoint.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "bos_token": "<s>",
31
+ "clean_up_tokenization_spaces": false,
32
+ "eos_token": "</s>",
33
+ "legacy": false,
34
+ "model_max_length": 1000000000000000019884624838656,
35
+ "pad_token": "</s>",
36
+ "sp_model_kwargs": {},
37
+ "tokenizer_class": "LlamaTokenizer",
38
+ "unk_token": "<unk>",
39
+ "use_default_system_prompt": true
40
+ }
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "sequelbox/SpellBlade",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 8192,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 28672,
14
+ "max_position_embeddings": 4096,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 64,
17
+ "num_hidden_layers": 80,
18
+ "num_key_value_heads": 8,
19
+ "pad_token_id": 0,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_scaling": null,
23
+ "rope_theta": 10000.0,
24
+ "tie_word_embeddings": false,
25
+ "torch_dtype": "float32",
26
+ "transformers_version": "4.36.2",
27
+ "use_cache": false,
28
+ "vocab_size": 32000
29
+ }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.36.2",
7
+ "use_cache": false
8
+ }
model-00001-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5352d1d26fb1a4187e6655674d3d7b9e65578dd4f88e05e3053008c53acfe878
3
+ size 4806739440
model-00002-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e28499b8d3d1d6db61e3385c3736a1d8faaf891a2fc2ce89b71739f3b23a551
3
+ size 4630578440
model-00003-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a69dced67f39f9f6c60816642904fba0e123102682df36a7703c4da3ed2b3478
3
+ size 4362142864
model-00004-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c06846f87b3828821532d315409db90421ec463a8fb3057a2e28d5ebff4b368b
3
+ size 4966188864
model-00005-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da16ef7dba13ec0c23aa5721c1c39a1793f0aac98a122942afb50bfd33ab0d12
3
+ size 4362142864
model-00006-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7af94011077c5ff5cb6e41b28360c9ddecb491db47380274a61e687cea70001
3
+ size 4362142864
model-00007-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54c6b0177452845eef4a53ff47787150687efe0e8ae05cf3bcdb45ae9b85ce95
3
+ size 4966188864
model-00008-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b60807d26c02d8c09f84aae1e630e4c298eec45b90bc8a9bb4ff21b9886a65f0
3
+ size 4362142880
model-00009-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20fafa021662573a48905ec08221dd85ed0181a0088055832fa95997f92cbed6
3
+ size 4362142872
model-00010-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d86186d95a9895d1c2942dedaa390c45b8749fa7dc816461ee45fc08c764b4ec
3
+ size 4966188880
model-00011-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70fa9d2b4d1922ed02f61b527528d56444692ec3fc136998eb06eeb344194016
3
+ size 4362142872
model-00012-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e696bdf1acd81583bd846d5355bd401876398bc3b1ba8977846a1125dcaf1cc6
3
+ size 4362142872
model-00013-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eed999502c6eeff2b1a5158b16f7db3d3121c957ec9143eea3552eef8562082
3
+ size 4966188880
model-00014-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ed50e80d456ce42bab05404bc5a13f36cd668a8acfd055cbe945647b79bdf69
3
+ size 4362142872
model-00015-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24534c62918587285be9533a4cde711864570ecc425350fbb0278a591f2c6393
3
+ size 4362142872
model-00016-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:417705ff101a6516ed3d3a018fe5968109e6505eaf88966233fc3485956a1efa
3
+ size 4966188880
model-00017-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50ac743e5437c9321360795d533c3767a4454ba373d0ab781970c5b67d871568
3
+ size 4362142872
model-00018-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35bb3eb4e4d049834b41635404827f9df27ab3c996aa9734a25da421c200c2a0
3
+ size 4362142872
model-00019-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1cedc43136ff86f562bc67c9088d54e9b8e495c0052f31e975dda03d715a0c1
3
+ size 4966188880
model-00020-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:692706f2b9e2c65accd5e8c2cafbb12447f9b97968c30ceb7645de7f70a5111f
3
+ size 4362142872
model-00021-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70f933252a5599e0388aeeaf126f2a3c14f9be83724ed8a08bad491511ec7269
3
+ size 4362142872
model-00022-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:454788a5f8ed3ca95d84b6fae013ab9832d517c1714239240e80016d7a1c4d15
3
+ size 4966188880
model-00023-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4aefb059c46a9998dfc942961bfeedc719e854ed1189c0b4102709acf297e539
3
+ size 4362142872
model-00024-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15de4c62eeee6d8c6c58039c17c9dc1d298b73e93838880c5d8242842b75b05f
3
+ size 4362142872
model-00025-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f8a8d5a10febcf110b98976ffcec9ca8cd17d1fda8d7046b6726b04728b412c
3
+ size 4966188880
model-00026-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4afd46ecebb697ade9f2a6d9323a635e43bbb1f3364c91fe3434162eaffdb8af
3
+ size 4362142872
model-00027-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c7c91daeae73cfdbcb3167604a4e002fb3803e4ed0e21a5c9cb1fde61688835
3
+ size 4362142872
model-00028-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:602add813be1febc8a4330796075ad18927aea3c5215dea0f976fd8a87cac61f
3
+ size 4966188880
model-00029-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9d2db54aecdf9bdf854cb640be346b4023ef2857abeb720297c59627a7ea8e2
3
+ size 4362142872
model-00030-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c49e0ba95fa93516bb09eb6d6d8cdc1c9c36d9f11ea24d64834e5a81817a41a9
3
+ size 4362142872
model-00031-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74afde4aa694a3f0e7568bfc5c2cc72d14e632fb0b8c1c0b52cc34131588d03e
3
+ size 4966188880
model-00032-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c553a79c4dc7527158d1ee8c63e6cfdfed7cb544247e14f8a262a869db928ac1
3
+ size 4362142872
model-00033-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e64f4c3e6c63b879ab412b1b10c81411e2a885deea883c4d1cc1392e6ad06ee6
3
+ size 4362142872
model-00034-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a7c8d6bf49315f4a15792aad78c45ad306f997fc870fa639e82c2411c361362
3
+ size 4966188880
model-00035-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d57fb6faaecc3f1ce6646bd3c1576712e7e00e82bebc099475710f1a30ed133e
3
+ size 4362142872
model-00036-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2543b012997bc2f260198166c51c328e98edd2e4a764317c43147aa6ee1834d6
3
+ size 4362142872
model-00037-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2caf2871528d1a8a5871d90208b53312e3b821f09ca8c9aa6ac7ef8e55fc6a4
3
+ size 4966188880
model-00038-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:818c322947e1675269b0e4513cc8cb8168849f46e9ce8c4bec12b57c6a5a3fbe
3
+ size 4362142872
model-00039-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0f4aae1fc83d6c5a7235e7255b5f86c28b42b1077d1237878b59bcc0ac53b86
3
+ size 4362142872
model-00040-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5ad4528388d48d8854250724ad0025e05091bb4af040186d0cd11aa508e7a90
3
+ size 4966188880
model-00041-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8bf95940073d1e160e5349dbacafff6aa9b93d56ee7dc17500462c38b913dec
3
+ size 4362142872
model-00042-of-00061.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1f30feb65fd41e48016368c5c12a077ae2c2c01cb2690bf465288c9141eefaa
3
+ size 4362142872