diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3a977cd172a732de23c6f2699dbb7bd22ab078f3 --- /dev/null +++ b/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "airev-ai/Jais-Inception-70b-V1.2.2", + "architectures": [ + "Qwen2ForCausalLM" + ], + "attention_dropout": 0.0, + "eos_token_id": 151645, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 29568, + "max_position_embeddings": 131072, + "max_window_layers": 28, + "model_type": "qwen2", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "rms_norm_eps": 1e-05, + "rope_theta": 1000000.0, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.43.3", + "use_cache": false, + "use_sliding_window": false, + "vocab_size": 152064 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..3ffdb66c1e17e6272ba24f39d0ee3474574c9d70 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,7 @@ +{ + "bos_token_id": 151643, + "do_sample": true, + "eos_token_id": 151643, + "max_new_tokens": 2048, + "transformers_version": "4.43.3" +} diff --git a/model-00001-of-00082.safetensors b/model-00001-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..efb86dd80e25853745d341271cc1403d7dcab9ff --- /dev/null +++ b/model-00001-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0176936410ae77d886b3e1ec011b0d7de259fd6f78113232a18375958855a090 +size 2491416712 diff --git a/model-00002-of-00082.safetensors b/model-00002-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d5fc2f6e16ff9248c994664af84b3cd4ca6abe10 --- /dev/null +++ b/model-00002-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03f75999aeaf65cd2e2a4b61c64455aa73666af565f80de56e0b6bb173664642 +size 1923164168 diff --git a/model-00003-of-00082.safetensors b/model-00003-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..066b756bd24f29eb5823275fc39cc322334f747f --- /dev/null +++ b/model-00003-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcb446777f5d16b252054d235cb99e83b75683a3e07ff858f299606ed929b10a +size 1889588696 diff --git a/model-00004-of-00082.safetensors b/model-00004-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f768955063a6c13ac2435a0c1e59831dbe8da520 --- /dev/null +++ b/model-00004-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4948766bce7ac96fa50160817d9e4398112c2acb48776ba2aa006fcf2749cd43 +size 1755370856 diff --git a/model-00005-of-00082.safetensors b/model-00005-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a838dc7b19156c0cf0236541fc5ba88cb6594050 --- /dev/null +++ b/model-00005-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0683e150f73bc92ff8d8d6b32fc02cf9f9955982018ab3e31727c7006debb5f4 +size 1755370856 diff --git a/model-00006-of-00082.safetensors b/model-00006-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c6edbc098ba5314f0eae8c8005ae5efca318f2e1 --- /dev/null +++ b/model-00006-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:239f7429b57ed5c6f3b451d3982f15bf361f8f67673ec73c5b925712b02f0c24 +size 1755370856 diff --git a/model-00007-of-00082.safetensors b/model-00007-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ed7b6a9181040194faf52542ed6072ec1daa5ef3 --- /dev/null +++ b/model-00007-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f56970349fbf55968af01554a3edd07aa271f7e395471ec87e7c6a2e8fae7e4 +size 1755370856 diff --git a/model-00008-of-00082.safetensors b/model-00008-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ec9122c5634be40ccb4ebbd58197b03953f9c17e --- /dev/null +++ b/model-00008-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4864e4a6ae846d9bde4aa7a1bf4b9dc41555f037eba47aa897afb883815614fc +size 1755370856 diff --git a/model-00009-of-00082.safetensors b/model-00009-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6ac684cdfbae9b46984b3787beef355f165a15ab --- /dev/null +++ b/model-00009-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd35d5a113fd5c6a3e6088846df6a01cf4b3cc616ebf305d4508852ac57303c9 +size 1755370856 diff --git a/model-00010-of-00082.safetensors b/model-00010-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b74ddaec8bfd23395d7b2bddfe6a8959ae68415f --- /dev/null +++ b/model-00010-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:beee3f524c7785459ab338b7733b62ed1b423bef5b15a08c8b7c255116bb2c01 +size 1755370856 diff --git a/model-00011-of-00082.safetensors b/model-00011-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..60ce7c01a5eafd96f0dcfee635edd1adaa56f217 --- /dev/null +++ b/model-00011-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6aa3797f7c0109ca71267e3b5908740b28b69168616682a3d68268dad5e073f0 +size 1755370840 diff --git a/model-00012-of-00082.safetensors b/model-00012-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..788e757488caf664a1afb31480559b7f1a7062cc --- /dev/null +++ b/model-00012-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf2537f883f37298a0792f12b35fa658b76962b4d5ac6954d4684ba8989ab120 +size 1755370864 diff --git a/model-00013-of-00082.safetensors b/model-00013-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4a64859098f98318ad73d3fb70a25c78f3dd1691 --- /dev/null +++ b/model-00013-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a412fe6b191da60d5ab1434c28a28f3c7b154886d71c717eae9942a7c5e63bd +size 1755370864 diff --git a/model-00014-of-00082.safetensors b/model-00014-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cce4bbf4036464b6d12aa659caf7e8fea79e591c --- /dev/null +++ b/model-00014-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c2bf73a0fb8003d26b9500cd5532751f443f38e5efcb2472a67fcb788c11e6b +size 1755370864 diff --git a/model-00015-of-00082.safetensors b/model-00015-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7e0a3ad9c319acfe5126b60d5001f7371a43f0a6 --- /dev/null +++ b/model-00015-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce122f1d632730839d0d7cf1e9185d69a92a76ee1adbfb838047900632761fe7 +size 1755370864 diff --git a/model-00016-of-00082.safetensors b/model-00016-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..031535a5b4f66545f8eda876aa6fcd60e41b5758 --- /dev/null +++ b/model-00016-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcd05c0b31c63fd09ea8c5f81b23df51123f3f2516167ac8bd381598d572b5e7 +size 1755370864 diff --git a/model-00017-of-00082.safetensors b/model-00017-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8724d6427ebdf6cdacad32354f31d45b6c12df0c --- /dev/null +++ b/model-00017-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6eecc64d4bf343797c43a6fea1646578e54abf9326914e83d69e18cbc0ed5a30 +size 1755370864 diff --git a/model-00018-of-00082.safetensors b/model-00018-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81bc584306cded06637679fe2b2fd91e5eaf4f06 --- /dev/null +++ b/model-00018-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1de3042cdc54d00be9a0cb80bb17a49d5d316c4ae2b9e3b2f46eb21dae58446f +size 1755370864 diff --git a/model-00019-of-00082.safetensors b/model-00019-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6b6591cf5fec35e456274f788e91fa3dcfc2b773 --- /dev/null +++ b/model-00019-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8eabe258c7a471e8c394ca95e50666d112a881b2dc40e29bf5b4c73a7bc904ad +size 1755370864 diff --git a/model-00020-of-00082.safetensors b/model-00020-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a71bf76bb3b01f974d967d84d474a42738d1de71 --- /dev/null +++ b/model-00020-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3488d93822951c63de6a1fbd1d2a28efb01ca84854d664f746fe2db6f6e2dba +size 1755370864 diff --git a/model-00021-of-00082.safetensors b/model-00021-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3f3669fb65d66b6b6626a3d60dcc7b98974070cc --- /dev/null +++ b/model-00021-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1aee5654f2f930e99b4f9ac8718851200e34e15bab534b3093aa187fca87f768 +size 1755370864 diff --git a/model-00022-of-00082.safetensors b/model-00022-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..80ab6cdf9960f2457dbeb1536f4b25aef40fcb84 --- /dev/null +++ b/model-00022-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:185e459515d7ac83e2905e826e3a6fad47e07ee9121218f6cd5418709abe8265 +size 1755370864 diff --git a/model-00023-of-00082.safetensors b/model-00023-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e18e809a1be2433cb684b3e75a14c432bf7c2969 --- /dev/null +++ b/model-00023-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:793204b76950a0949bd167627ca5500f88d913bbeafd9ce7645a70a9dc80f41f +size 1755370864 diff --git a/model-00024-of-00082.safetensors b/model-00024-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7dd5fa2163cfe1b288f969f6e676f2c9802e50aa --- /dev/null +++ b/model-00024-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01476917c43b0655e1c7be13573ec9a71102610f24b2bb4a4943dc586b8050ef +size 1755370864 diff --git a/model-00025-of-00082.safetensors b/model-00025-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da91ddc48a3384f453523bcbd4d0b7e2716ea3af --- /dev/null +++ b/model-00025-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8583d8c5358e8df3560b2b1936e901dad7ea17eda376d398924c466a7de3f94 +size 1755370864 diff --git a/model-00026-of-00082.safetensors b/model-00026-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3bd5735413ce26c8a36af90985b6e2113a5facad --- /dev/null +++ b/model-00026-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc0beab5d8b835275f0178313760bd0f79073347931fb6cd1d23e7a965fcb122 +size 1755370864 diff --git a/model-00027-of-00082.safetensors b/model-00027-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8154d920ac4b08c9d3762d3dbbf292bf2c8c04e2 --- /dev/null +++ b/model-00027-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d51272f1e35e1d5ab7d24173272dd4b68ada3a6f0f9627ea897ac562810c8934 +size 1755370864 diff --git a/model-00028-of-00082.safetensors b/model-00028-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e7d0e978e27103ebf08f8e6d6ba4823e21b14572 --- /dev/null +++ b/model-00028-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d61b35d7a2ea41356de108b4927f0a69d593aebad5b4923c622f351d982e7a3 +size 1755370864 diff --git a/model-00029-of-00082.safetensors b/model-00029-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..79a04bebe9f4c32529c7a25fb1432edbb2934b92 --- /dev/null +++ b/model-00029-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec7099ce3aab622cc8d72b5b2be4af1f8d52291020f28ca4286802e74be38ce5 +size 1755370864 diff --git a/model-00030-of-00082.safetensors b/model-00030-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0f1004bdd79c1350067078991980fb7a6032274c --- /dev/null +++ b/model-00030-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5967c993cdb19770f1a7ef9227c0553f0684faeea3dfffdfd6d0fd6ca922efe +size 1755370864 diff --git a/model-00031-of-00082.safetensors b/model-00031-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b97e7b3f979a1753c6dda7fb00338dd1e52418df --- /dev/null +++ b/model-00031-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acc8d1760fbdcf0b9cfcbd042460d21c798e93b6e3b664974bc0ade10f472f90 +size 1755370864 diff --git a/model-00032-of-00082.safetensors b/model-00032-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e9157ae377cfa6355c80d8ad2933afa824f90231 --- /dev/null +++ b/model-00032-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ea248ef5601e2ade2f125d4a335134f3359bc86796b7ff8f0ffc64ff9a22aec +size 1755370864 diff --git a/model-00033-of-00082.safetensors b/model-00033-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a48bc670b9745e8781c7770901fe6bf524f4e5f3 --- /dev/null +++ b/model-00033-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fff1bf1fcb2c3f67d5a01d0ec15efb56baf708ec0f2867c3157ddfbfd17d20e2 +size 1755370864 diff --git a/model-00034-of-00082.safetensors b/model-00034-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ba5bad04419a9836c9710cc7f4dfe68bf0a199c6 --- /dev/null +++ b/model-00034-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e97cfdfd5062d2abb4ba472b119504c491e26232020893a800c64a66ab7bb965 +size 1755370864 diff --git a/model-00035-of-00082.safetensors b/model-00035-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a115a20f28646e963ad6b66067b843fed73c5d9d --- /dev/null +++ b/model-00035-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3498feb73011a09602f0c723f91c07a5a1263ff8621921dce81acd72cf46e493 +size 1755370864 diff --git a/model-00036-of-00082.safetensors b/model-00036-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..346cfbbdded336aecf5dd967dd9a8ead31f078ae --- /dev/null +++ b/model-00036-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a1ceaa07c15b36a8696785049bfba9ad467957977b6c3a2bb1b82df2751a90f +size 1755370864 diff --git a/model-00037-of-00082.safetensors b/model-00037-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dd55165b4adebb8c9da8817039e34b1d0e7cc543 --- /dev/null +++ b/model-00037-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e35f32f2de587fc593d7bb28baec20ad68e7c65be10be8af4f4eea1397f1e5a +size 1755370864 diff --git a/model-00038-of-00082.safetensors b/model-00038-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..124d608187a423329ef944b91f36607663ec349b --- /dev/null +++ b/model-00038-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be8552ab946fd46181a0fe675c49ec6c9992adbe5ff6bdb35ffd2852d66adfd1 +size 1755370864 diff --git a/model-00039-of-00082.safetensors b/model-00039-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..775c7b99673382ae50a4bfa30f09b2279c1d1027 --- /dev/null +++ b/model-00039-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a96c7d0c3d322caa5ae44cdb167fb7aaad4489c16d4cd5c2947425fd2a485f7 +size 1755370864 diff --git a/model-00040-of-00082.safetensors b/model-00040-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fa74c86fa414e6e98582dc1373a01fc8f34aa5b8 --- /dev/null +++ b/model-00040-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:842651fb16e0b2f6aaed4aff713611fbbbf3dc54f95fb7279ba775b0a29114f1 +size 1755370864 diff --git a/model-00041-of-00082.safetensors b/model-00041-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..112f48c4eddb9b4e0345615a2251da2072a91332 --- /dev/null +++ b/model-00041-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25341543ea72cc79f551d7ae590aef6a901828580faae64b6ad833be96702085 +size 1755370864 diff --git a/model-00042-of-00082.safetensors b/model-00042-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..33e91a9673d19438750673bbb602622ea303117d --- /dev/null +++ b/model-00042-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f2d093093e1f16efda816c8fa6191afad3f6b07304ae7ccaaf13a9d5bb18c2f +size 1755370864 diff --git a/model-00043-of-00082.safetensors b/model-00043-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..44071a80a5510b9edcb4e01d4e59bd5f24e3f2e7 --- /dev/null +++ b/model-00043-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fac6e0d017e28dab039ce4513934573d668112c2722f0b4a7f4236317755bfd +size 1755370864 diff --git a/model-00044-of-00082.safetensors b/model-00044-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b65e6a345f58d628c752912b798d56014d02bb2c --- /dev/null +++ b/model-00044-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a03ee2828fb95f13b311be09c6bc7da64b11fb6cbef30d25789d600c8c364aad +size 1755370864 diff --git a/model-00045-of-00082.safetensors b/model-00045-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f0c08b50589906c2b85496851dcb919679f0c166 --- /dev/null +++ b/model-00045-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdd2c8cc96870d62bdb03bbbca0735319e65d62c7c42dc7bbc72d411ad533415 +size 1755370864 diff --git a/model-00046-of-00082.safetensors b/model-00046-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a576a24557e6ff652dde278d6e3de7937a86e545 --- /dev/null +++ b/model-00046-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1456ebc2ff34e9ef4928dd4a14e0b43a085b68b60eb9615ad56b84ef0d91a7de +size 1755370864 diff --git a/model-00047-of-00082.safetensors b/model-00047-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..271172219061766e42d6c2e97513d3b0a87331ab --- /dev/null +++ b/model-00047-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f1e1d403784a685f7b40adc2bbe07eaef7053e8391ff427a63f03199a1d2f73 +size 1755370864 diff --git a/model-00048-of-00082.safetensors b/model-00048-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f89923192f189865ed46178e40d52dd3ad872e0f --- /dev/null +++ b/model-00048-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30ae4eb23a9a34dac7884c6fb5f644f5d3dc65110d5e68861574942df84a91d4 +size 1755370864 diff --git a/model-00049-of-00082.safetensors b/model-00049-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6d929afc623c60ab70b06fc7f96bdc0300f3a889 --- /dev/null +++ b/model-00049-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13d24ced26458e5ddb3e872138562965701bf9ff378dceef088ee9c5d6e8d15e +size 1755370864 diff --git a/model-00050-of-00082.safetensors b/model-00050-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..de7ae854dcdd8bdc94e9cad96556b0a5ad5d55b1 --- /dev/null +++ b/model-00050-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fde40dd48e0a009ed7c7530abfd5ad8c259991eb49a19ac6d23d1ae3f6ff16f0 +size 1755370864 diff --git a/model-00051-of-00082.safetensors b/model-00051-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3c7ca5e0f344b0814cd69369e7d5f93f141cf86a --- /dev/null +++ b/model-00051-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d90c802dc63928fc0a486d0a786d5bc7dfe064cf284ef7803a0390948d74e487 +size 1755370864 diff --git a/model-00052-of-00082.safetensors b/model-00052-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e7cf31457c124597d02e69e06d2b497b81393b13 --- /dev/null +++ b/model-00052-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bb814fd1c163af130c727ebeeb484d7207128da1ec6692450463c3f7028f6f6 +size 1755370864 diff --git a/model-00053-of-00082.safetensors b/model-00053-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..640bab954310ffa0d4c862b459038acc321762ea --- /dev/null +++ b/model-00053-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a87c66fd0d5c8b60d7f77b408857e04175c576614a1084dbafca6da986225ea +size 1755370864 diff --git a/model-00054-of-00082.safetensors b/model-00054-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..936b3e626d391709d0e577e6287b10c9f1996851 --- /dev/null +++ b/model-00054-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d201b331f70c95e4c6684b4c68a76588d55e00a683439ccdf52fd89d38eecfd3 +size 1755370864 diff --git a/model-00055-of-00082.safetensors b/model-00055-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..56be27e34b72bc7f1f96e95b1c62cee0d2421cd3 --- /dev/null +++ b/model-00055-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8229a427b13ac194cd959ffc8040fc0e963332257b964673e162e9c55bceaf9 +size 1755370864 diff --git a/model-00056-of-00082.safetensors b/model-00056-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81154247060f5bc8b1e4859607cdddb9c5fb338f --- /dev/null +++ b/model-00056-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5aa056417d811b6e5b97c0dbc438378f8dd8c7cae961a5a326cdc6091c306633 +size 1755370864 diff --git a/model-00057-of-00082.safetensors b/model-00057-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..229020d4b8fd71b58c5e84f93acc4699973d0a57 --- /dev/null +++ b/model-00057-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3225c90d3a03e718a40bf28aa9016f8b94e7cb9f1bfac949742b1a0bef5b2c8b +size 1755370864 diff --git a/model-00058-of-00082.safetensors b/model-00058-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ca66842527041bb246043318eda3f0fe209c7825 --- /dev/null +++ b/model-00058-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4575e4e56ace6829b270e2c21d35aee59c4c7e55a72c41afc4bf094e177d8c02 +size 1755370864 diff --git a/model-00059-of-00082.safetensors b/model-00059-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..00b8eebf791d24841fa462b57d1335cd262e26fe --- /dev/null +++ b/model-00059-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70fa9018db895bfc23efba2f494f35fde8d1d0958504665d6146763207a85d8c +size 1755370864 diff --git a/model-00060-of-00082.safetensors b/model-00060-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..58ffef6deab88c93535abae73c91167bb3d0478f --- /dev/null +++ b/model-00060-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f45a0642a9197ee7724770677a9572a5773fff3930af54927c199fb3f9ce698f +size 1755370864 diff --git a/model-00061-of-00082.safetensors b/model-00061-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..79c260c7b2816492e4432c7c18075e55d9c21724 --- /dev/null +++ b/model-00061-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29c41de77fc7d738dc09654d8df7474b026e082a2da2290c1807b7350bf6c2c9 +size 1755370864 diff --git a/model-00062-of-00082.safetensors b/model-00062-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..27edf5c34e269610cd3a34c9132d813af6987c9a --- /dev/null +++ b/model-00062-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a187027398a4ae16fcd0ec46f50cc8e96d0a3c2b43b9fb89bd990a618b89d73a +size 1755370864 diff --git a/model-00063-of-00082.safetensors b/model-00063-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..759bfde08f01d5404a271d64c22983b077d9a960 --- /dev/null +++ b/model-00063-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec18e80d5f8d11532c158c1a8352df30d9493864372ada6fd4e48ac53df4721b +size 1755370864 diff --git a/model-00064-of-00082.safetensors b/model-00064-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..96e5080a798a133125124bc786d33ad1a3be7935 --- /dev/null +++ b/model-00064-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6ea6bca46568057ca51ebabf2d7ebbf67a16f0af5403af22addad1016eaf583 +size 1755370864 diff --git a/model-00065-of-00082.safetensors b/model-00065-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e4ca41a78dd7fdde5c174f58ab54e358bfc59e46 --- /dev/null +++ b/model-00065-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26f55a22dc108719ecb770e880c1098305924272d020303bc966f5717795de66 +size 1755370864 diff --git a/model-00066-of-00082.safetensors b/model-00066-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c3ee552bba3e778b7ba177c40ac06ba2f82d2727 --- /dev/null +++ b/model-00066-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4834c139177c2e97a9081453556e798d2c0b97f10fc082228e1b96169d3dfcb0 +size 1755370864 diff --git a/model-00067-of-00082.safetensors b/model-00067-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4ab3d4abcea185d4374326cc88dfcb21cf7d4c2a --- /dev/null +++ b/model-00067-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:555cce966a241a442ad4904d0728d08c3a87250bd5c5d02f082f7208d74128f9 +size 1755370864 diff --git a/model-00068-of-00082.safetensors b/model-00068-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a6cc6b86575a0304dc5a6fbbaf4107d76897bab9 --- /dev/null +++ b/model-00068-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ff5656d171f64d3cfc0d18d7f77c4d32b2ecd8661c61dd9d2f77c4763448020 +size 1755370864 diff --git a/model-00069-of-00082.safetensors b/model-00069-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..12f1b747949adba39be89304d3b6e038e2277635 --- /dev/null +++ b/model-00069-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d344b5127a82c6c148d67212f87c6c17251346c167acc79f6b1bb29d64205c55 +size 1755370864 diff --git a/model-00070-of-00082.safetensors b/model-00070-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f0106af4f6138d90d34ec9bb8d2c053d18389014 --- /dev/null +++ b/model-00070-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b41a4727240db8257af0a074587da33039973a06838f3df1ec654c9b852c9a6 +size 1755370864 diff --git a/model-00071-of-00082.safetensors b/model-00071-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d6fa6dd867c9aba8a131e0e517f9936f4706424b --- /dev/null +++ b/model-00071-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5230d6fd458cc2dc7ac51541387f5f9efab9dde4056a36d077e2ed8022725069 +size 1755370864 diff --git a/model-00072-of-00082.safetensors b/model-00072-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a620f9ff18f565570f218c8d2530d86047ab6a90 --- /dev/null +++ b/model-00072-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f3b7e5643a1f2eeeb139dd2a35cdac56431450b02804f0c6ecb97033c51dbdb +size 1755370864 diff --git a/model-00073-of-00082.safetensors b/model-00073-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..80bbe12e2a5223c4a063145141f147f24b6473ee --- /dev/null +++ b/model-00073-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91e7bd35cacf16200732d23ef0c42bc69a07cb9701980a91b2461b53329c7282 +size 1755370864 diff --git a/model-00074-of-00082.safetensors b/model-00074-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b3cdd1149dbf41bebc2e5f0a2cf26c4cb8816f7c --- /dev/null +++ b/model-00074-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0af36250ceeedad3c23a722bb5a03c1d3e62ffa2278d6e7431f39f81cbb074a +size 1755370864 diff --git a/model-00075-of-00082.safetensors b/model-00075-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..700c25390248d5cca8a4da404f40d0924f235887 --- /dev/null +++ b/model-00075-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:759ffc8b39e43d3071f3abf86f8a9c411c1f42d5aa31d738cbe4a20defa1329a +size 1755370864 diff --git a/model-00076-of-00082.safetensors b/model-00076-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fb10d4e1bd837c33c55498d8606a1e7ce019c069 --- /dev/null +++ b/model-00076-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f886d49329a4aab306e630e24b389e81dcf2047a4475357e3810c73507450d2e +size 1755370864 diff --git a/model-00077-of-00082.safetensors b/model-00077-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..37bb34e17f618ccce2f3ef981c6961e977461114 --- /dev/null +++ b/model-00077-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc9d0c375e372e246b1cbef3a92268acf90fe31951c90d00b9c78b23380c21d5 +size 1755370864 diff --git a/model-00078-of-00082.safetensors b/model-00078-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8a7472a89ab0915583668622e2243c151605be51 --- /dev/null +++ b/model-00078-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11450d48e65256fd5603814a6338c7a2073379fe015888da0523383edcb274b1 +size 1755370864 diff --git a/model-00079-of-00082.safetensors b/model-00079-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8851f8e45db72227651b049f7efbe82329ff9b65 --- /dev/null +++ b/model-00079-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec3768b3e10dd83250fc2674137f9b79e8003cf6255a90edd44fc781afe452df +size 1755370864 diff --git a/model-00080-of-00082.safetensors b/model-00080-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53157207d3f1ffb531f5b0a10088ea980c94abe7 --- /dev/null +++ b/model-00080-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:004eb216c64febb5355daa6f36ea28b50a2d35e6e2b884527fc16357e1d2c988 +size 1755370864 diff --git a/model-00081-of-00082.safetensors b/model-00081-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e46d5eab863aa63c5397da210830611fca5de4b2 --- /dev/null +++ b/model-00081-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1a5741e940d8038c2ed5988b810e9e7e06328bd5fc88995d213e742c6430011 +size 2491416704 diff --git a/model-00082-of-00082.safetensors b/model-00082-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..88f949478f7c7282dd45e524bf20b211c9c66d61 --- /dev/null +++ b/model-00082-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:666b212cfa708aa708203a968b817183dcccfe27a9822d93ebe8cfa26a9cff45 +size 1453376176 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..7586a1e894d26d2be9b5cf5a2e1201e6f61d2b7a --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,970 @@ +{ + "metadata": { + "total_size": 145412407296 + }, + "weight_map": { + "lm_head.weight": "model-00081-of-00082.safetensors", + "model.embed_tokens.weight": "model-00001-of-00082.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00082.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.k_proj.bias": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.q_proj.bias": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.v_proj.bias": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00082.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00082.safetensors", + "model.layers.1.self_attn.k_proj.bias": "model-00002-of-00082.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.self_attn.q_proj.bias": "model-00002-of-00082.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.1.self_attn.v_proj.bias": "model-00002-of-00082.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.10.input_layernorm.weight": "model-00012-of-00082.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00012-of-00082.safetensors", + "model.layers.10.self_attn.k_proj.bias": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.q_proj.bias": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.v_proj.bias": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.11.input_layernorm.weight": "model-00013-of-00082.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00013-of-00082.safetensors", + "model.layers.11.self_attn.k_proj.bias": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.q_proj.bias": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.v_proj.bias": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.12.input_layernorm.weight": "model-00014-of-00082.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00014-of-00082.safetensors", + "model.layers.12.self_attn.k_proj.bias": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.q_proj.bias": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.v_proj.bias": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.13.input_layernorm.weight": "model-00015-of-00082.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00015-of-00082.safetensors", + "model.layers.13.self_attn.k_proj.bias": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.q_proj.bias": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.v_proj.bias": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.14.input_layernorm.weight": "model-00016-of-00082.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00016-of-00082.safetensors", + "model.layers.14.self_attn.k_proj.bias": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.q_proj.bias": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.v_proj.bias": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.15.input_layernorm.weight": "model-00017-of-00082.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00017-of-00082.safetensors", + "model.layers.15.self_attn.k_proj.bias": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.q_proj.bias": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.v_proj.bias": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.16.input_layernorm.weight": "model-00018-of-00082.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00018-of-00082.safetensors", + "model.layers.16.self_attn.k_proj.bias": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.q_proj.bias": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.v_proj.bias": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.17.input_layernorm.weight": "model-00019-of-00082.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00019-of-00082.safetensors", + "model.layers.17.self_attn.k_proj.bias": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.q_proj.bias": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.v_proj.bias": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.18.input_layernorm.weight": "model-00020-of-00082.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00020-of-00082.safetensors", + "model.layers.18.self_attn.k_proj.bias": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.q_proj.bias": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.v_proj.bias": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.19.input_layernorm.weight": "model-00021-of-00082.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00021-of-00082.safetensors", + "model.layers.19.self_attn.k_proj.bias": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.q_proj.bias": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.v_proj.bias": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.2.input_layernorm.weight": "model-00004-of-00082.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00082.safetensors", + "model.layers.2.self_attn.k_proj.bias": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.q_proj.bias": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.v_proj.bias": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.20.input_layernorm.weight": "model-00022-of-00082.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00022-of-00082.safetensors", + "model.layers.20.self_attn.k_proj.bias": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.q_proj.bias": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.v_proj.bias": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.21.input_layernorm.weight": "model-00023-of-00082.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00023-of-00082.safetensors", + "model.layers.21.self_attn.k_proj.bias": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.q_proj.bias": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.v_proj.bias": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.22.input_layernorm.weight": "model-00024-of-00082.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00024-of-00082.safetensors", + "model.layers.22.self_attn.k_proj.bias": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.q_proj.bias": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.v_proj.bias": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.23.input_layernorm.weight": "model-00025-of-00082.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00025-of-00082.safetensors", + "model.layers.23.self_attn.k_proj.bias": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.q_proj.bias": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.v_proj.bias": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.24.input_layernorm.weight": "model-00026-of-00082.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00026-of-00082.safetensors", + "model.layers.24.self_attn.k_proj.bias": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.q_proj.bias": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.v_proj.bias": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.25.input_layernorm.weight": "model-00027-of-00082.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00027-of-00082.safetensors", + "model.layers.25.self_attn.k_proj.bias": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.q_proj.bias": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.v_proj.bias": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.26.input_layernorm.weight": "model-00028-of-00082.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00028-of-00082.safetensors", + "model.layers.26.self_attn.k_proj.bias": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.q_proj.bias": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.v_proj.bias": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.27.input_layernorm.weight": "model-00029-of-00082.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00029-of-00082.safetensors", + "model.layers.27.self_attn.k_proj.bias": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.q_proj.bias": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.v_proj.bias": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.28.input_layernorm.weight": "model-00030-of-00082.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00030-of-00082.safetensors", + "model.layers.28.self_attn.k_proj.bias": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.q_proj.bias": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.v_proj.bias": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.29.input_layernorm.weight": "model-00031-of-00082.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00031-of-00082.safetensors", + "model.layers.29.self_attn.k_proj.bias": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.q_proj.bias": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.v_proj.bias": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.3.input_layernorm.weight": "model-00005-of-00082.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00005-of-00082.safetensors", + "model.layers.3.self_attn.k_proj.bias": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.q_proj.bias": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.v_proj.bias": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.30.input_layernorm.weight": "model-00032-of-00082.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00032-of-00082.safetensors", + "model.layers.30.self_attn.k_proj.bias": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.q_proj.bias": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.v_proj.bias": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.31.input_layernorm.weight": "model-00033-of-00082.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00033-of-00082.safetensors", + "model.layers.31.self_attn.k_proj.bias": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.q_proj.bias": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.v_proj.bias": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.32.input_layernorm.weight": "model-00034-of-00082.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00034-of-00082.safetensors", + "model.layers.32.self_attn.k_proj.bias": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.q_proj.bias": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.v_proj.bias": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.33.input_layernorm.weight": "model-00035-of-00082.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00035-of-00082.safetensors", + "model.layers.33.self_attn.k_proj.bias": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.q_proj.bias": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.v_proj.bias": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.34.input_layernorm.weight": "model-00036-of-00082.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00036-of-00082.safetensors", + "model.layers.34.self_attn.k_proj.bias": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.q_proj.bias": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.v_proj.bias": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.35.input_layernorm.weight": "model-00037-of-00082.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00037-of-00082.safetensors", + "model.layers.35.self_attn.k_proj.bias": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.q_proj.bias": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.v_proj.bias": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.36.input_layernorm.weight": "model-00038-of-00082.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00038-of-00082.safetensors", + "model.layers.36.self_attn.k_proj.bias": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.q_proj.bias": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.v_proj.bias": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.37.input_layernorm.weight": "model-00039-of-00082.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00039-of-00082.safetensors", + "model.layers.37.self_attn.k_proj.bias": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.q_proj.bias": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.v_proj.bias": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.38.input_layernorm.weight": "model-00040-of-00082.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00040-of-00082.safetensors", + "model.layers.38.self_attn.k_proj.bias": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.q_proj.bias": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.v_proj.bias": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.39.input_layernorm.weight": "model-00041-of-00082.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00041-of-00082.safetensors", + "model.layers.39.self_attn.k_proj.bias": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.q_proj.bias": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.v_proj.bias": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.4.input_layernorm.weight": "model-00006-of-00082.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00006-of-00082.safetensors", + "model.layers.4.self_attn.k_proj.bias": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.q_proj.bias": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.v_proj.bias": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.40.input_layernorm.weight": "model-00042-of-00082.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00042-of-00082.safetensors", + "model.layers.40.self_attn.k_proj.bias": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.q_proj.bias": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.v_proj.bias": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.41.input_layernorm.weight": "model-00043-of-00082.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00043-of-00082.safetensors", + "model.layers.41.self_attn.k_proj.bias": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.q_proj.bias": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.v_proj.bias": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.42.input_layernorm.weight": "model-00044-of-00082.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00044-of-00082.safetensors", + "model.layers.42.self_attn.k_proj.bias": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.q_proj.bias": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.v_proj.bias": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.43.input_layernorm.weight": "model-00045-of-00082.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00045-of-00082.safetensors", + "model.layers.43.self_attn.k_proj.bias": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.q_proj.bias": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.v_proj.bias": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.44.input_layernorm.weight": "model-00046-of-00082.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00046-of-00082.safetensors", + "model.layers.44.self_attn.k_proj.bias": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.q_proj.bias": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.v_proj.bias": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.45.input_layernorm.weight": "model-00047-of-00082.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00047-of-00082.safetensors", + "model.layers.45.self_attn.k_proj.bias": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.q_proj.bias": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.v_proj.bias": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.46.input_layernorm.weight": "model-00048-of-00082.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00048-of-00082.safetensors", + "model.layers.46.self_attn.k_proj.bias": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.q_proj.bias": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.v_proj.bias": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.47.input_layernorm.weight": "model-00049-of-00082.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00049-of-00082.safetensors", + "model.layers.47.self_attn.k_proj.bias": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.q_proj.bias": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.v_proj.bias": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.48.input_layernorm.weight": "model-00050-of-00082.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00050-of-00082.safetensors", + "model.layers.48.self_attn.k_proj.bias": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.q_proj.bias": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.v_proj.bias": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.49.input_layernorm.weight": "model-00051-of-00082.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00051-of-00082.safetensors", + "model.layers.49.self_attn.k_proj.bias": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.q_proj.bias": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.v_proj.bias": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.5.input_layernorm.weight": "model-00007-of-00082.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00007-of-00082.safetensors", + "model.layers.5.self_attn.k_proj.bias": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.q_proj.bias": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.v_proj.bias": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.50.input_layernorm.weight": "model-00052-of-00082.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00052-of-00082.safetensors", + "model.layers.50.self_attn.k_proj.bias": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.q_proj.bias": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.v_proj.bias": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.51.input_layernorm.weight": "model-00053-of-00082.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00053-of-00082.safetensors", + "model.layers.51.self_attn.k_proj.bias": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.q_proj.bias": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.v_proj.bias": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.52.input_layernorm.weight": "model-00054-of-00082.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00054-of-00082.safetensors", + "model.layers.52.self_attn.k_proj.bias": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.q_proj.bias": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.v_proj.bias": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.53.input_layernorm.weight": "model-00055-of-00082.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00055-of-00082.safetensors", + "model.layers.53.self_attn.k_proj.bias": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.q_proj.bias": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.v_proj.bias": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.54.input_layernorm.weight": "model-00056-of-00082.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00056-of-00082.safetensors", + "model.layers.54.self_attn.k_proj.bias": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.q_proj.bias": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.v_proj.bias": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.55.input_layernorm.weight": "model-00057-of-00082.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00057-of-00082.safetensors", + "model.layers.55.self_attn.k_proj.bias": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.q_proj.bias": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.v_proj.bias": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.56.input_layernorm.weight": "model-00058-of-00082.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00058-of-00082.safetensors", + "model.layers.56.self_attn.k_proj.bias": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.q_proj.bias": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.v_proj.bias": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.57.input_layernorm.weight": "model-00059-of-00082.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00059-of-00082.safetensors", + "model.layers.57.self_attn.k_proj.bias": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.q_proj.bias": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.v_proj.bias": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.58.input_layernorm.weight": "model-00060-of-00082.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00060-of-00082.safetensors", + "model.layers.58.self_attn.k_proj.bias": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.q_proj.bias": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.v_proj.bias": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.59.input_layernorm.weight": "model-00061-of-00082.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00061-of-00082.safetensors", + "model.layers.59.self_attn.k_proj.bias": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.q_proj.bias": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.v_proj.bias": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.6.input_layernorm.weight": "model-00008-of-00082.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00008-of-00082.safetensors", + "model.layers.6.self_attn.k_proj.bias": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.q_proj.bias": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.v_proj.bias": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.60.input_layernorm.weight": "model-00062-of-00082.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00062-of-00082.safetensors", + "model.layers.60.self_attn.k_proj.bias": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.q_proj.bias": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.v_proj.bias": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.61.input_layernorm.weight": "model-00063-of-00082.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00063-of-00082.safetensors", + "model.layers.61.self_attn.k_proj.bias": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.q_proj.bias": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.v_proj.bias": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.62.input_layernorm.weight": "model-00064-of-00082.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00064-of-00082.safetensors", + "model.layers.62.self_attn.k_proj.bias": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.q_proj.bias": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.v_proj.bias": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.63.input_layernorm.weight": "model-00065-of-00082.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00065-of-00082.safetensors", + "model.layers.63.self_attn.k_proj.bias": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.q_proj.bias": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.v_proj.bias": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.64.input_layernorm.weight": "model-00066-of-00082.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00066-of-00082.safetensors", + "model.layers.64.self_attn.k_proj.bias": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.q_proj.bias": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.v_proj.bias": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.65.input_layernorm.weight": "model-00067-of-00082.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00067-of-00082.safetensors", + "model.layers.65.self_attn.k_proj.bias": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.q_proj.bias": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.v_proj.bias": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.66.input_layernorm.weight": "model-00068-of-00082.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00068-of-00082.safetensors", + "model.layers.66.self_attn.k_proj.bias": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.q_proj.bias": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.v_proj.bias": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.67.input_layernorm.weight": "model-00069-of-00082.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00069-of-00082.safetensors", + "model.layers.67.self_attn.k_proj.bias": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.q_proj.bias": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.v_proj.bias": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.68.input_layernorm.weight": "model-00070-of-00082.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00070-of-00082.safetensors", + "model.layers.68.self_attn.k_proj.bias": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.q_proj.bias": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.v_proj.bias": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.69.input_layernorm.weight": "model-00071-of-00082.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00071-of-00082.safetensors", + "model.layers.69.self_attn.k_proj.bias": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.q_proj.bias": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.v_proj.bias": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.7.input_layernorm.weight": "model-00009-of-00082.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00009-of-00082.safetensors", + "model.layers.7.self_attn.k_proj.bias": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.q_proj.bias": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.v_proj.bias": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.70.input_layernorm.weight": "model-00072-of-00082.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00072-of-00082.safetensors", + "model.layers.70.self_attn.k_proj.bias": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.q_proj.bias": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.v_proj.bias": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.71.input_layernorm.weight": "model-00073-of-00082.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00073-of-00082.safetensors", + "model.layers.71.self_attn.k_proj.bias": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.q_proj.bias": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.v_proj.bias": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.72.input_layernorm.weight": "model-00074-of-00082.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00074-of-00082.safetensors", + "model.layers.72.self_attn.k_proj.bias": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.q_proj.bias": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.v_proj.bias": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.73.input_layernorm.weight": "model-00075-of-00082.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00075-of-00082.safetensors", + "model.layers.73.self_attn.k_proj.bias": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.q_proj.bias": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.v_proj.bias": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.74.input_layernorm.weight": "model-00076-of-00082.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00076-of-00082.safetensors", + "model.layers.74.self_attn.k_proj.bias": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.q_proj.bias": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.v_proj.bias": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.75.input_layernorm.weight": "model-00077-of-00082.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00077-of-00082.safetensors", + "model.layers.75.self_attn.k_proj.bias": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.q_proj.bias": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.v_proj.bias": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.76.input_layernorm.weight": "model-00078-of-00082.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00078-of-00082.safetensors", + "model.layers.76.self_attn.k_proj.bias": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.q_proj.bias": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.v_proj.bias": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.77.input_layernorm.weight": "model-00079-of-00082.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00079-of-00082.safetensors", + "model.layers.77.self_attn.k_proj.bias": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.q_proj.bias": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.v_proj.bias": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.78.input_layernorm.weight": "model-00080-of-00082.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00080-of-00082.safetensors", + "model.layers.78.self_attn.k_proj.bias": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.q_proj.bias": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.v_proj.bias": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.79.input_layernorm.weight": "model-00082-of-00082.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00082-of-00082.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00082-of-00082.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00082-of-00082.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00082-of-00082.safetensors", + "model.layers.79.self_attn.k_proj.bias": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.q_proj.bias": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.v_proj.bias": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.8.input_layernorm.weight": "model-00010-of-00082.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00010-of-00082.safetensors", + "model.layers.8.self_attn.k_proj.bias": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.q_proj.bias": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.v_proj.bias": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.9.input_layernorm.weight": "model-00011-of-00082.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00011-of-00082.safetensors", + "model.layers.9.self_attn.k_proj.bias": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.q_proj.bias": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.v_proj.bias": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00010-of-00082.safetensors", + "model.norm.weight": "model-00082-of-00082.safetensors" + } +}