diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..cacd7c317b49e2ff5c8009b286ccab8ed385251f --- /dev/null +++ b/README.md @@ -0,0 +1,201 @@ +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3faf27bebce68b58bcda91fdf6d056da3a58450a --- /dev/null +++ b/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "/workspace/Llama2-70B/FT", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 4096, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 10000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.38.0.dev0", + "use_cache": true, + "vocab_size": 32000 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..232ce1d0095899ee851870118e550e973aeb3b50 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,10 @@ +{ + "bos_token_id": 1, + "do_sample": true, + "eos_token_id": 2, + "max_length": 4096, + "pad_token_id": 0, + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.38.0.dev0" +} diff --git a/model-00001-of-00061.safetensors b/model-00001-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..88226faa73b1a57743730a42957b4ef7c7a5216e --- /dev/null +++ b/model-00001-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd53bffad672c1aecb14dd65d12f007e12ebc2228f99fc030f02bf3e0404b791 +size 4806739440 diff --git a/model-00002-of-00061.safetensors b/model-00002-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..277bb2c433ba66f71986e651dace290f5d5e8579 --- /dev/null +++ b/model-00002-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36b0ed2a3afb55ed7461c313704a0b191747fd20e7ff09acbb0cb9d5423e8940 +size 4630578440 diff --git a/model-00003-of-00061.safetensors b/model-00003-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bd98ab9eedfbf10deb805cbfd1c2e5175357a2fd --- /dev/null +++ b/model-00003-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9dc02a30b95c017a6cc478350afe9543fc16c2248f84d6f2880724f48a15936a +size 4362142864 diff --git a/model-00004-of-00061.safetensors b/model-00004-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eafdaa8cbcc155c610b4a079ec7c8258a7443900 --- /dev/null +++ b/model-00004-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c283d8f6511d90796546cd2251cca239da3e3a2372b99d7ff00da2d00594302 +size 4966188864 diff --git a/model-00005-of-00061.safetensors b/model-00005-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..58143d753cb89e3a5fc8f861a08c2ddd023b2917 --- /dev/null +++ b/model-00005-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1071231035a3995ef6139c05a9a47be2d68dc04aca89d6368f594bf377adba6 +size 4362142864 diff --git a/model-00006-of-00061.safetensors b/model-00006-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fcc9f3806a4ff6ac686149dbac1f02c37c2e1126 --- /dev/null +++ b/model-00006-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49abc698f3034d2fca827c1c9143a13258e408148fa49fe18ff64e07255292e6 +size 4362142864 diff --git a/model-00007-of-00061.safetensors b/model-00007-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a62246c13a5d51db011adecd2c1a797bc60bce34 --- /dev/null +++ b/model-00007-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:690ae5401bb4b9f2e1fa2b19b42f6265536e773de0539d594a6a781a1f4f28aa +size 4966188864 diff --git a/model-00008-of-00061.safetensors b/model-00008-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..49d1e3aff440abb0db12f4f6853d9f1b21907bb1 --- /dev/null +++ b/model-00008-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58a97a878fa2767f7dee26baeaf5e48703696936b1a9a7e9cf3254423382313e +size 4362142880 diff --git a/model-00009-of-00061.safetensors b/model-00009-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fb2b017bf6825f9faa73c4aced6da34746fff54c --- /dev/null +++ b/model-00009-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59360c6feaac0c0a03f6d6685333a3d3a69741f3f98eba49181cd814dd5a9f43 +size 4362142872 diff --git a/model-00010-of-00061.safetensors b/model-00010-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6c086d251f5f9645438d2140c018f2c7157eacaa --- /dev/null +++ b/model-00010-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a133fbfa5279d861a98fc2dd87388770ecfc176e81f4e441c8fac714938ebec +size 4966188880 diff --git a/model-00011-of-00061.safetensors b/model-00011-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e28443240ac57d81157774f74afe35f660ae313b --- /dev/null +++ b/model-00011-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79dacf1cda9f65cb515eb26c4126223393e7c49f341252c2b57db13ccfda7335 +size 4362142872 diff --git a/model-00012-of-00061.safetensors b/model-00012-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b8130e8f79e6a03023bfbd2d6ea682af58c76e0c --- /dev/null +++ b/model-00012-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39b2cdfe6f3b1f919bd6b1ce77b277a14ff1934e89e18945c1b123b7de93a255 +size 4362142872 diff --git a/model-00013-of-00061.safetensors b/model-00013-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a6379c370f69e060b3acfbf137676862b701599f --- /dev/null +++ b/model-00013-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d280b5f4f9d1b730716624a4875c41651e288958b012cb49c8e68185c5d4ce51 +size 4966188880 diff --git a/model-00014-of-00061.safetensors b/model-00014-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1f619e0d45fca20e3bc1b041012778f1267e02d5 --- /dev/null +++ b/model-00014-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d82adaa8880d1196324a2fea9985809d1f1d9dfa900c0e0239ec9349ef76f3d7 +size 4362142872 diff --git a/model-00015-of-00061.safetensors b/model-00015-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5fd78f306219bbc8c9420e0f446b002063a93f5d --- /dev/null +++ b/model-00015-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a58a98b625befafcb67e92f4e564b009fb53a046d0dcb29bf8e009a0d2cdd77 +size 4362142872 diff --git a/model-00016-of-00061.safetensors b/model-00016-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4406d1d8213e34dbfbf541103bb1f676f28dc17d --- /dev/null +++ b/model-00016-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:710f7e8db78a2a39574bbab16ddf420f7a282f7fd98171c7ee5cb4e47eadeaba +size 4966188880 diff --git a/model-00017-of-00061.safetensors b/model-00017-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b1a7614c2bbcd090f6b98573db03079d7875b8cc --- /dev/null +++ b/model-00017-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6923c9478d2f98c49be7c9d1be3d2d7aefdb655e47cae04572d9ca245a59c44 +size 4362142872 diff --git a/model-00018-of-00061.safetensors b/model-00018-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..40a30d92f6cbc9009e9880ab458724a5a32445ed --- /dev/null +++ b/model-00018-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74640537d0369b08ca8a469d5dc29e9eeba32c4c4915cfcd622cde79bc86d2ef +size 4362142872 diff --git a/model-00019-of-00061.safetensors b/model-00019-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ddc0d777e460c438584e49fe1ae5c6c184271584 --- /dev/null +++ b/model-00019-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1ba6eea04282c5892e1e1189f7070b300993662f54a1da2da5476ef6a8d3571 +size 4966188880 diff --git a/model-00020-of-00061.safetensors b/model-00020-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4d0588744e936a8a4703b955ad5d6ae9469f0f48 --- /dev/null +++ b/model-00020-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:154040c12792a18f04ab72b5bf8fff865646fe44df418e778fab12665c94b189 +size 4362142872 diff --git a/model-00021-of-00061.safetensors b/model-00021-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..de23a60c0d3634ded052658078c82093326cdfa1 --- /dev/null +++ b/model-00021-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54d3aaa7ab2d26de3e04a04d35baea3b3f25efa0897f669cde512ba6657128c8 +size 4362142872 diff --git a/model-00022-of-00061.safetensors b/model-00022-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..402b6047796f060d437b7c0183f015335c905358 --- /dev/null +++ b/model-00022-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bbd71741ceab9eb9608cceb962933e21efccc1e53924b2c1bc571952b0c1d987 +size 4966188880 diff --git a/model-00023-of-00061.safetensors b/model-00023-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..113d2036c68e1fa509a5f0cc8bf5931b95803a1f --- /dev/null +++ b/model-00023-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e00c3d61501e309bcd2857b5803e620fe29504ee61cf50d6e77095c8535642e +size 4362142872 diff --git a/model-00024-of-00061.safetensors b/model-00024-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e4cb0b596ef28d1d5630ed19d9c685a623ab8e98 --- /dev/null +++ b/model-00024-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f7f75f906c8c3246ded63c35bf31a074a6772fc78127d0c8e49f8411bfa0161 +size 4362142872 diff --git a/model-00025-of-00061.safetensors b/model-00025-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4c874122f85ff7035be7df1a8fab1401b3fc666 --- /dev/null +++ b/model-00025-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5defb2c0e777cd9eeaafb015720a3e5d1899e48303630b2ec0c19a4a7c9b57c +size 4966188880 diff --git a/model-00026-of-00061.safetensors b/model-00026-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c4712a264d88825a8347e5856a94f0148bcd7ccc --- /dev/null +++ b/model-00026-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:715e786f0884d984f94ed5580b010c8738522d0a45da0995e0b5858e949ee95d +size 4362142872 diff --git a/model-00027-of-00061.safetensors b/model-00027-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8305d0812a45ae619806f3fc6d19a87cc338ce36 --- /dev/null +++ b/model-00027-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84965ad6ebcb9fc6eea0ab1f404093a83cb4e82821e5232534721b50e1f850dc +size 4362142872 diff --git a/model-00028-of-00061.safetensors b/model-00028-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..90ced294143c50d7dd404b2f8e98c26877223e78 --- /dev/null +++ b/model-00028-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d3d9f74c123a1aa1787f4bfb3dc59e2e86b7bde787a2f5ced8240e05f80f840 +size 4966188880 diff --git a/model-00029-of-00061.safetensors b/model-00029-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2161176b217d2486340969a42bcd0aaa6617c2b5 --- /dev/null +++ b/model-00029-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76291c9ec8b1dacc7542f7d8cc31e1fd59b6434e2dfac414fabedcb44247edef +size 4362142872 diff --git a/model-00030-of-00061.safetensors b/model-00030-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fd38eec03662f5244676869197c5f1abd475b044 --- /dev/null +++ b/model-00030-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1442eeba11be283bc49d515ea56974f23158f604bd3ad219bc72a7956bf970dc +size 4362142872 diff --git a/model-00031-of-00061.safetensors b/model-00031-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..16d16d34f512f1078b96a100e34b11f3e5a3085d --- /dev/null +++ b/model-00031-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b5304c0c26144f4c1d6e27bc7b141aba877832d00cb0bbc8f7233ff6ddc29d1 +size 4966188880 diff --git a/model-00032-of-00061.safetensors b/model-00032-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..28b3d432fcfbc03a5dca3108fbfbd59dcb623968 --- /dev/null +++ b/model-00032-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3de1a7489f79fde0cc257ee05fb02d0e2f6fb652f849c3f4048f76c4be3be7bc +size 4362142872 diff --git a/model-00033-of-00061.safetensors b/model-00033-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4f1c7350901650d80f259fa1ee2377a6097f7a7 --- /dev/null +++ b/model-00033-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66ecc053ec32e8b7ca313ec57e0037bad94aed8e028b6db24f7ea7c53bfbef7e +size 4362142872 diff --git a/model-00034-of-00061.safetensors b/model-00034-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1e895db657eab37a1991466fcfa96d4f48366c66 --- /dev/null +++ b/model-00034-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a621f36108588990c5502eba2621978afd88fbec270e339c426d6d3b7d354add +size 4966188880 diff --git a/model-00035-of-00061.safetensors b/model-00035-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d2e757431255ec85a9422a293d75f5cb5dcfd5fb --- /dev/null +++ b/model-00035-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51c2c6123f0ecdbcaec9bcedd045129f7dc51a162a0d581895ce50bbaf7186d1 +size 4362142872 diff --git a/model-00036-of-00061.safetensors b/model-00036-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..69da724315635eef0defbe2a32274376d81d99e9 --- /dev/null +++ b/model-00036-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a47824541f0482149205c1e5a11febdb12772f676f78a293b5ff2d2e55b1d5a0 +size 4362142872 diff --git a/model-00037-of-00061.safetensors b/model-00037-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..48260de35078f04773eb88fca78ac103c51d04ab --- /dev/null +++ b/model-00037-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8cc9f9de9abb20875cba0f25e3180854e5177e1ab23a03759445e2fe36e608e2 +size 4966188880 diff --git a/model-00038-of-00061.safetensors b/model-00038-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..21fdf86eb8b4dc67639261fcae1a8f8efc054503 --- /dev/null +++ b/model-00038-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a6fd2cc8cc400ee5a027018a2248ea8dab4c5128bc4cc5faa9d547ed95bff37 +size 4362142872 diff --git a/model-00039-of-00061.safetensors b/model-00039-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ac6e6c8e5c99cf9f85f1a945b49d15a51d0042db --- /dev/null +++ b/model-00039-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5924a38338075132298090288fba3b3cce9f494b73837b78d78ce9ddebb7c698 +size 4362142872 diff --git a/model-00040-of-00061.safetensors b/model-00040-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..36bca3c432757eea525600a24e3d4be16c69273d --- /dev/null +++ b/model-00040-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:143cb228263733142287e5e1ea7dcc0786730391f21110dd4b4eb3ffb7b67e98 +size 4966188880 diff --git a/model-00041-of-00061.safetensors b/model-00041-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..922c5b04f5a05ed79372d917b56cbfee965ab84e --- /dev/null +++ b/model-00041-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62cc5088bfb59aec834bec23dd449e6326f3f605a85dc2214ec10b47a8660889 +size 4362142872 diff --git a/model-00042-of-00061.safetensors b/model-00042-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8e342e072502c33ad4c8d960f5b6f1503865bd96 --- /dev/null +++ b/model-00042-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95ee09baf15ad92d93e3f91ee9235692397a48aa8c266eec1b5025138bc1eff6 +size 4362142872 diff --git a/model-00043-of-00061.safetensors b/model-00043-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..593035e6318f51b9aa3ea887363f143197930dde --- /dev/null +++ b/model-00043-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b559df52bce9eb952d69740362114f9e8c210fd22ac38ab0659acfc80e4104 +size 4966188880 diff --git a/model-00044-of-00061.safetensors b/model-00044-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..80712d5982eb9662155cbad921b7c10e471a6570 --- /dev/null +++ b/model-00044-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6f5ebb4c57ee2a08540844024234e24df70f1b81743ce56488f4ef566b0f8b3 +size 4362142872 diff --git a/model-00045-of-00061.safetensors b/model-00045-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..36e70d5d7f3be12ffa6e49a10daf82810f8b8fbf --- /dev/null +++ b/model-00045-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa0cdb72f3bcb91d4c6a4f2f6c8ec76600d22b9e3c47f3d757d9a0013c7caea1 +size 4362142872 diff --git a/model-00046-of-00061.safetensors b/model-00046-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..67e4ffb57a4d198512037b605d6e23107b91f323 --- /dev/null +++ b/model-00046-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bb8acbf2db88813241e8321b99cc1a5bb51cd1ec83afba5e314ddbd7d578745 +size 4966188880 diff --git a/model-00047-of-00061.safetensors b/model-00047-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bbf5c8f09ef7b87944d8be59ba48ccc986a0c07f --- /dev/null +++ b/model-00047-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44355fbb7bfd019f6f2bb2b8539ebb3ac538ca5db6639192997e9e2d6db35495 +size 4362142872 diff --git a/model-00048-of-00061.safetensors b/model-00048-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fe95b22532f5becb72d741d9ad4b54af3fe9db44 --- /dev/null +++ b/model-00048-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e19bb6f4e7acda450f1943d11523c94fc719729343626a51772bdfaba6ca319c +size 4362142872 diff --git a/model-00049-of-00061.safetensors b/model-00049-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ac5fbe39b7dc66cf7698c9ac2771137728a11aa8 --- /dev/null +++ b/model-00049-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7c6b3d2c01f51b54974458b796b0b8dcbab00b05544ca1c85fcc6751c32e350 +size 4966188880 diff --git a/model-00050-of-00061.safetensors b/model-00050-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..805ec12f218a91819edf86da053256e64e159294 --- /dev/null +++ b/model-00050-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3153fdba177ac166114190c534a0389cf8715cbc177f3e9678b43f97d8539c0 +size 4362142872 diff --git a/model-00051-of-00061.safetensors b/model-00051-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a8d8082e5c181c30f530dd4aa4091d78e769ae78 --- /dev/null +++ b/model-00051-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dd15a89bad99cc861ff691356f9fde81ba9a2dba3a4ead94c248633d90fd2cc +size 4362142872 diff --git a/model-00052-of-00061.safetensors b/model-00052-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a2a796755058b7048c4e09f4e9256136bde294ea --- /dev/null +++ b/model-00052-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0403ebe6ce81478feee5cc83b42672608d41f2ffe2be0ce35bf293b8a998ebb +size 4966188880 diff --git a/model-00053-of-00061.safetensors b/model-00053-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53ca523d437cf2553acd085623380646378d4e3c --- /dev/null +++ b/model-00053-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85c15a95968f4bb93e9c3bc757af82a65056dfb4510b895aee84e94822dccebc +size 4362142872 diff --git a/model-00054-of-00061.safetensors b/model-00054-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bc53ab0ed99edd73f56d196a366cab74ab1ec4d7 --- /dev/null +++ b/model-00054-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68c3f2e532abe742bba6056d49ca4c99e443709e22e3e975fee08a0c7404e350 +size 4362142872 diff --git a/model-00055-of-00061.safetensors b/model-00055-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..67b0b508ebff1350b5311d5bd5f21e9037e37749 --- /dev/null +++ b/model-00055-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:160ecffd201280e9e9d10ac489cdc3e8b2527d1e54161d7a9d75482490e6a7ba +size 4966188880 diff --git a/model-00056-of-00061.safetensors b/model-00056-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2bf49dc134c8169033e10d58e17f6ae1debf30f2 --- /dev/null +++ b/model-00056-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19a773face767239bae6298109d794e64cc65d2bd4fa796900c51d12c84d8b48 +size 4362142872 diff --git a/model-00057-of-00061.safetensors b/model-00057-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..baf0f3f304c7edecd2c21c7dd87685f1863e4246 --- /dev/null +++ b/model-00057-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e75548cee9608371e79e7e7e3d7ab886d5d0b7605688433ec35efc54840744e +size 4362142872 diff --git a/model-00058-of-00061.safetensors b/model-00058-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..acf2c88681a38709592d4900303e11ba755f35a3 --- /dev/null +++ b/model-00058-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a92cbd16e6495fb2e47a8b219856c151e9eb26eeb80942daa5646bf41cebc45b +size 4966188880 diff --git a/model-00059-of-00061.safetensors b/model-00059-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2cbc2dbbb2c3d4550e9cf90915dfb799cea12e28 --- /dev/null +++ b/model-00059-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8d12e54be243d51571b323ace7ac2409740b6d01a9201743019c98115039465 +size 4362142872 diff --git a/model-00060-of-00061.safetensors b/model-00060-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fbc445ccecc0fd7da12573746e6c8fb1df5a3325 --- /dev/null +++ b/model-00060-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1df8a758e8e8ed3011223225a8ce24ed3d95d47d19b3e29c7b1331e15ccd0bf2 +size 4362142872 diff --git a/model-00061-of-00061.safetensors b/model-00061-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1d048079c2e3fda364ff3978fadf34e8ceff963a --- /dev/null +++ b/model-00061-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f211436028040e1988b1f088ff96737441b9fd043055cdd5d6de9248fbdaa75 +size 1988198960 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..4156aa9dfeed60097cda66ec60af33ad1607de57 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 275906592768 + }, + "weight_map": { + "lm_head.weight": "model-00061-of-00061.safetensors", + "model.embed_tokens.weight": "model-00001-of-00061.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.10.input_layernorm.weight": "model-00009-of-00061.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00061.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.12.input_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.input_layernorm.weight": "model-00011-of-00061.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00061.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.14.input_layernorm.weight": "model-00012-of-00061.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00061.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.16.input_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.input_layernorm.weight": "model-00014-of-00061.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00061.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.18.input_layernorm.weight": "model-00015-of-00061.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00061.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00061.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00061.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.20.input_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.input_layernorm.weight": "model-00017-of-00061.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00061.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.22.input_layernorm.weight": "model-00018-of-00061.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00061.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.24.input_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.input_layernorm.weight": "model-00020-of-00061.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00061.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.26.input_layernorm.weight": "model-00021-of-00061.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00061.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.28.input_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.input_layernorm.weight": "model-00023-of-00061.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00061.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.30.input_layernorm.weight": "model-00024-of-00061.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00061.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.32.input_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.input_layernorm.weight": "model-00026-of-00061.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00061.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.34.input_layernorm.weight": "model-00027-of-00061.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00061.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.36.input_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.input_layernorm.weight": "model-00029-of-00061.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00061.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.38.input_layernorm.weight": "model-00030-of-00061.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00061.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.40.input_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.input_layernorm.weight": "model-00032-of-00061.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00061.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.42.input_layernorm.weight": "model-00033-of-00061.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00061.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.44.input_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.input_layernorm.weight": "model-00035-of-00061.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00061.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.46.input_layernorm.weight": "model-00036-of-00061.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00061.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.48.input_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.input_layernorm.weight": "model-00038-of-00061.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00061.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00061.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00061.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.50.input_layernorm.weight": "model-00039-of-00061.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00061.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.52.input_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.input_layernorm.weight": "model-00041-of-00061.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00061.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.54.input_layernorm.weight": "model-00042-of-00061.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00061.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.56.input_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.input_layernorm.weight": "model-00044-of-00061.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00061.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.58.input_layernorm.weight": "model-00045-of-00061.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00061.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.6.input_layernorm.weight": "model-00006-of-00061.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00061.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.60.input_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.input_layernorm.weight": "model-00047-of-00061.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00061.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.62.input_layernorm.weight": "model-00048-of-00061.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00061.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.64.input_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.input_layernorm.weight": "model-00050-of-00061.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00061.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.66.input_layernorm.weight": "model-00051-of-00061.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00061.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.68.input_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.input_layernorm.weight": "model-00053-of-00061.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00061.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.70.input_layernorm.weight": "model-00054-of-00061.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00061.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.72.input_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.input_layernorm.weight": "model-00056-of-00061.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00061.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.74.input_layernorm.weight": "model-00057-of-00061.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00061.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.76.input_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.input_layernorm.weight": "model-00059-of-00061.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00061.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.78.input_layernorm.weight": "model-00060-of-00061.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00061.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00061.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00061.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00061.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.8.input_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.input_layernorm.weight": "model-00008-of-00061.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00061.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", + "model.norm.weight": "model-00061-of-00061.safetensors" + } +}