MikeRoz commited on
Commit
6aee357
1 Parent(s): 00b8712

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.1
3
+ base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
4
+ pipeline_tag: text-generation
5
+ ---
6
+ # Reflection Llama-3.1 70B
7
+
8
+ **Reflection 70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.**
9
+
10
+ The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
11
+
12
+ You can [try the model here](https://reflection-playground-production.up.railway.app/).
13
+
14
+ ## Benchmarks
15
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/60518f3731c5be7f3dd5ebc3/zNs-ZFs0SbnomH7mikiOU.png)
16
+
17
+ All benchmarks tested have been checked for contamination by running [LMSys's LLM Decontaminator](https://github.com/lm-sys/llm-decontaminator). When benchmarking, we isolate the `<output>` and benchmark on solely that section.
18
+
19
+ Trained from Llama 3.1 70B Instruct, you can sample from Reflection 70B using the same code, pipelines, etc. as any other Llama model. It even uses the stock Llama 3.1 chat template format (though, we've trained in a few new special tokens to aid in reasoning and reflection).
20
+
21
+ During sampling, the model will start by outputting reasoning inside `<thinking>` and `</thinking>` tags, and then once it is satisfied with its reasoning, it will output the final answer inside `<output>` and `</output>` tags. Each of these tags are special tokens, trained into the model.
22
+
23
+ This enables the model to separate its internal thoughts and reasoning from its final answer, improving the experience for the user.
24
+
25
+ Inside the `<thinking>` section, the model may output one or more `<reflection>` tags, which signals the model has caught an error in its reasoning and will attempt to correct it before providing a final answer.
26
+
27
+ ## System Prompt
28
+
29
+ ```
30
+ The system prompt used for training this model is:
31
+
32
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.
33
+
34
+ We recommend using this exact system prompt to get the best results from Reflection 70B. You may also want to experiment combining this system prompt with your own custom instructions to customize the behavior of the model.
35
+ ```
36
+
37
+ ## Chat Format
38
+
39
+ As mentioned above, the model uses the standard Llama 3.1 chat format. Here’s an example:
40
+
41
+ ```
42
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
43
+
44
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
45
+
46
+ what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
47
+ ```
48
+
49
+ ## Tips for Performance
50
+
51
+ - We are initially recommending a `temperature` of `.7` and a `top_p` of `.95`.
52
+ - For increased accuracy, append `Think carefully.` at the end of your messages.
53
+
54
+ ## Dataset / Report
55
+
56
+ Both the dataset and a brief report detailing how we trained this model will be released next week, alongside our Reflection 405B model that we expect will be the top-performing LLM in the world, including closed-source models.
57
+
58
+ ---
59
+
60
+ Thanks to Jason Kuperberg and Josh Bickett from the [HyperWrite](https://hyperwriteai.com) team for reviewing drafts of the report we'll be releasing next week.
61
+
62
+ Also, we know right now the model is split into a ton of files. We'll condense this soon to make the model easier to download and work with!
config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "meta-llama/Meta-Llama-3.1-70B-Instruct",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 128000,
9
+ "eos_token_id": [
10
+ 128001,
11
+ 128008,
12
+ 128009
13
+ ],
14
+ "hidden_act": "silu",
15
+ "hidden_size": 8192,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 28672,
18
+ "max_position_embeddings": 131072,
19
+ "mlp_bias": false,
20
+ "model_type": "llama",
21
+ "num_attention_heads": 64,
22
+ "num_hidden_layers": 80,
23
+ "num_key_value_heads": 8,
24
+ "pretraining_tp": 1,
25
+ "rms_norm_eps": 1e-05,
26
+ "rope_scaling": {
27
+ "factor": 8.0,
28
+ "high_freq_factor": 4.0,
29
+ "low_freq_factor": 1.0,
30
+ "original_max_position_embeddings": 8192,
31
+ "rope_type": "llama3"
32
+ },
33
+ "rope_theta": 500000.0,
34
+ "tie_word_embeddings": false,
35
+ "torch_dtype": "float32",
36
+ "transformers_version": "4.44.0",
37
+ "use_cache": true,
38
+ "vocab_size": 128256,
39
+ "quantization_config": {
40
+ "quant_method": "exl2",
41
+ "version": "0.1.9",
42
+ "bits": 8.0,
43
+ "head_bits": 8,
44
+ "calibration": {
45
+ "rows": 115,
46
+ "length": 2048,
47
+ "dataset": "(default)"
48
+ }
49
+ }
50
+ }
generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 128000,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 128001,
6
+ 128008,
7
+ 128009
8
+ ],
9
+ "temperature": 0.7,
10
+ "top_p": 0.95,
11
+ "transformers_version": "4.44.0"
12
+ }
huggingface-metadata.txt ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ url: https://huggingface.co/mattshumer/Reflection-70B
2
+ branch: main
3
+ download date: 2024-09-05 21:02:56
4
+ sha256sum:
5
+ 6527624b1c58c3fd50a271bc055e6fc4cba414ff09e34176f796e49256794429 model-00001-of-00162.safetensors
6
+ e9af83540da2b85a3a694ffc8706b5d5e2718b36668852cf8f9bdb91d65bcf6b model-00002-of-00162.safetensors
7
+ 44ce6875e055a1d3fe1485f6be750a53973f2b662e37c7453cb9383a915d1586 model-00003-of-00162.safetensors
8
+ 5f3a98e947923c7a784329410d89b7eac3b592ac668e755684fea98e23136aa5 model-00004-of-00162.safetensors
9
+ fc808eafe8e5951247a6cf701bf60546d0671634b1acb7624e252b03a6e6f4ee model-00005-of-00162.safetensors
10
+ 69ebc96ca1314e1ea3b3228d2a50e7dddeed36d4e1ef4eba52ea0aeaac4e4147 model-00006-of-00162.safetensors
11
+ 766a8c842b953e0bd7c8fc605b1397265823b177d26ce9bdbf2604d1ad746237 model-00007-of-00162.safetensors
12
+ 241e21a958223c4cd7c76972630a0900da9f0ce8ffb1666db8f2201d23851751 model-00008-of-00162.safetensors
13
+ ce152de02aec802c8560453a24a77c8dd8004f218ea011934f74b3dcb32b02f0 model-00009-of-00162.safetensors
14
+ 09b5c931e69c332054fd44b005f92e2533ad4aaac5e9d8981c6e5b083870df16 model-00010-of-00162.safetensors
15
+ d59a1dcc5f17be68d4ea3db1a1f9ec724b004591aea84de3981c83af860e8cc1 model-00011-of-00162.safetensors
16
+ 3df622381af2eaeb70965a909916053a6c90ea342ae86e08497c9f0f9c547a9f model-00012-of-00162.safetensors
17
+ 8bf702d935292db235264ae4f3f78c98053baac2ce6eeef0b66b846a3015344f model-00013-of-00162.safetensors
18
+ 7bb9cd37e7bc87565d4147eda9f83b6e158483b81b4bcd096a20b938c7943de8 model-00014-of-00162.safetensors
19
+ bfa9e6a3f2138b71b57c993fab2acff3d48ec282db8cb940403294b418230a75 model-00015-of-00162.safetensors
20
+ bde36db20468fbf887f672fa855b4800b6066daae217571d316026381b061169 model-00016-of-00162.safetensors
21
+ d4091dd2578fa59e2e19551dc9495eb91996a11d8d6960f14e2967b83c53898a model-00017-of-00162.safetensors
22
+ 6df521b397f7f9256da3d5956939ec6a584438d32de7da93f18907da680b8339 model-00018-of-00162.safetensors
23
+ 477375e767804aa122eac211b8c11525dde58b17041ac14e8972468dbfa44121 model-00019-of-00162.safetensors
24
+ 4bd87896352e70078d671544465fe955b871ab5f3419ad33051bdc86f5d1337a model-00020-of-00162.safetensors
25
+ 2f95bd73dc05e6f4367db50bc8ab8e9c080e01235158d339c78ed32c0e5d04c4 model-00021-of-00162.safetensors
26
+ 40dea034cccba996c020c0a2d22e4685b93675208af91d11ee65bfc0be5b48c5 model-00022-of-00162.safetensors
27
+ a61301bbed7ac1d0f0062f7d7250d578a86a0c00b35e83da6012ff5c90682345 model-00023-of-00162.safetensors
28
+ aaa155cdc6be42b55c758a2a2cacbdb0563f92c2474511a33edda9e3f15123c6 model-00024-of-00162.safetensors
29
+ f1f5e3b8831d43170f0f5d1fbf988dda8efad77bd6509af7cf9499a38c006b0b model-00025-of-00162.safetensors
30
+ 01d2fb287d88ceaa7d257ef0a4e645f8651e7b88898f6f45fb1cfea8af94e61a model-00026-of-00162.safetensors
31
+ 4db45a3256c99cdcfe006001eacafd0e0c1a6ba9bbc48ea975fb9fd7a7471cbf model-00027-of-00162.safetensors
32
+ 096dd9d39a8f56f3545399d58394b0079c938930409135e85a71c16c429c1e8a model-00028-of-00162.safetensors
33
+ 19d193fa0048415b9e254cb888bf5a8b2549fddfc4934c95e05eb3d70591595f model-00029-of-00162.safetensors
34
+ 90f99159cdf690aa8e8a388536a583c439afea525aea1ef73b7317fae0f1af53 model-00030-of-00162.safetensors
35
+ 0569735f89997cada6cbae4dfc3fe39d23b817fa8497045cccb7831e6295d1f2 model-00031-of-00162.safetensors
36
+ 9fcda19e75581bf3fcc77ea2afe1cb8e1262ba9effa6ed7a3af1bd11d35316d1 model-00032-of-00162.safetensors
37
+ 6738784c8d0461d519763af6613700ebdb72566beb4a7a31f480afda718e051a model-00033-of-00162.safetensors
38
+ 1eddeee1ef7a4bc4d4101dff08ae6abd1ade29c05fad24726570ca826b22b4a0 model-00034-of-00162.safetensors
39
+ a781cc7ef0ab3d5fcb1380a948f0570ba0fe9984ed36cbae15f67141f76868f6 model-00035-of-00162.safetensors
40
+ 39b5356f7477e8ee76b4c1431d0f9f431e6e3c54f66dc0ba7ee7d30a8feea116 model-00036-of-00162.safetensors
41
+ d8e2ade44644276466cd61c608690bffe96e279e5417fca207fa2cbf61f35f82 model-00037-of-00162.safetensors
42
+ 59c33507806ed231e43e604dfe8b93c491ea689b23fb4861554dc062dd56e4f7 model-00038-of-00162.safetensors
43
+ ea6002e4d8a68211434d1a72376a4ef9597b6cdf03ed792d5a505d7ad63d4e63 model-00039-of-00162.safetensors
44
+ e260421de937758f9755d16ecefb4e0fc68f8e53a006ed3199ad42de5c62eed5 model-00040-of-00162.safetensors
45
+ 1e159bfcd8cfd7abf33fe312781dc1047b4b99158eea31eab5a1adb4c1662450 model-00041-of-00162.safetensors
46
+ 3385db38cba9301a6fbe0b0bf654adfb68d866975ccbb11c1602337b234ce246 model-00042-of-00162.safetensors
47
+ 0d5bef48b1b28ccdca2d740d2b7563455d0c591592e0e5f6c5716c139ed69790 model-00043-of-00162.safetensors
48
+ 1ebca7e948332fe24f24612fbe6e7b894a9d5e1574b916ab4b37a412edba077a model-00044-of-00162.safetensors
49
+ 0351bd91c979c5a979d1878e40e11639fd3819d7fe808e8a18cebf5b671375dc model-00045-of-00162.safetensors
50
+ c7ae48d761ca02225dd062c21d1664e7fa73cd6f17bddd017f434e017d06b067 model-00046-of-00162.safetensors
51
+ 98b70aba868440ece4845976cbe60f3d05c023c64e443faaaf6c0cdb5a7600eb model-00047-of-00162.safetensors
52
+ 69e4369cfda195e65c5cf773f5778eab028ff1e1384ae1a4d6f308245d7a4654 model-00048-of-00162.safetensors
53
+ cc5828c98e43d4d176e5bc7da9d278860478e3d5d13e30d6c332f37a8dab1f89 model-00049-of-00162.safetensors
54
+ eed5f9d448b85f1c2483f1403979db58c1f66b25c30baf57f07659e33c5dd6a1 model-00050-of-00162.safetensors
55
+ ca53d034340ebeaf12d67017efcf5b6c71f41c291a946833bc8838a333ecd5a2 model-00051-of-00162.safetensors
56
+ 0c46ce6b01fe188f858c90ecd4d19b3caa131ae7cd10e20f386242f4f53ee4fd model-00052-of-00162.safetensors
57
+ 369048bf8b0ef6e1596e4e11f01835095b993c811a53ba18f36387980535eac2 model-00053-of-00162.safetensors
58
+ 32b829e2fa3da2b407a0fc99f74be93bf46a170743eb12a8120a9bc53729ca81 model-00054-of-00162.safetensors
59
+ 3ceb4a356f4d4e3b47ecb0b9a7eddd5a51d34127c6df3cdd28c0f3d24e37a83a model-00055-of-00162.safetensors
60
+ ecf07611c113bbf979960f2cb688a961b4dbe5dc86c57d2359650d83067ef288 model-00056-of-00162.safetensors
61
+ c8698532687a0d5919681dcd59dc2cb8846804e5fc0ee711befa5f159cb23f01 model-00057-of-00162.safetensors
62
+ b2d5fc525dde11e44a8ebc80f1a006be61536eff3b56fdcf8c90c6b0d9f25b58 model-00058-of-00162.safetensors
63
+ d58355fc0ff53771bd8be9626590ef008a894165436b8af27863910922265de1 model-00059-of-00162.safetensors
64
+ d4baebcf52b741c593e3a951b53dbc99a181f10ce7203a7480f94b8c0f05bd1d model-00060-of-00162.safetensors
65
+ 8958a93cf16ad284f559004b1709e1da0fefefabdf974b50f6695ed8fedf0ccc model-00061-of-00162.safetensors
66
+ 14969a1180727583b7eb1013e55563b7ad7065a0c69df186c4958144583c2d98 model-00062-of-00162.safetensors
67
+ b2119463c14ed76a0f087e574f7a9c3936c91393f7e52103bca25636826d161a model-00063-of-00162.safetensors
68
+ 311db4dfde5098de92e11dee91075807a03c02843e95a651e52813f69e622129 model-00064-of-00162.safetensors
69
+ ab9f67c753c40223c6cc32aa006a016a0107cb03d5a11ea25b26154e7855d446 model-00065-of-00162.safetensors
70
+ db7b7be268652ef1ed50c8f842d27f4c797ef59ad40461272486026a5b89a681 model-00066-of-00162.safetensors
71
+ 8190f91645a9c913ed3f8e9540da266f9758a42455906a65c855ffad00ab55ac model-00067-of-00162.safetensors
72
+ 130260d35ca06d58a6b69cdd01ed3ee78a337af11885ae8bcdb4a1cf6052273e model-00068-of-00162.safetensors
73
+ 1f244cef333e9b3bbef41af2128592f44803f31e1b0f09cf931c62f4c97234ff model-00069-of-00162.safetensors
74
+ a649f30fd533fc9b9a432f017973682740f0bd25702764f359b7881f24becab8 model-00070-of-00162.safetensors
75
+ 3cea57db736d36a57e5593c0170806d2aa37e48522f60b72a31cc3b8403d14cb model-00071-of-00162.safetensors
76
+ 84ff60e17e935532a826bf89bd976d0595545ff330ab9341e0d3f81e1618c541 model-00072-of-00162.safetensors
77
+ 7dafc2bddfc9e35d3a9a443ab540e7b2643d829261a0df72632baf9e6db6d700 model-00073-of-00162.safetensors
78
+ be53131fe155f5b6adf1d28b4773baf02ffb884c5ac2ab55058bcc94d9006668 model-00074-of-00162.safetensors
79
+ d07871d8ab4ce2a408de5f3e3f0875e31a0535e675903eab78a089ddd04b1383 model-00075-of-00162.safetensors
80
+ 8e52f072cc1038f7643de98a59360afefb1d5c1f13dcbd65bab179bb43bbaaa4 model-00076-of-00162.safetensors
81
+ a69161acc7964049bf5e91df1d734aef2380a812f8911f3cf6d54bc9f5c04f59 model-00077-of-00162.safetensors
82
+ 139a1907e243862eb774ce7eef2d37b157659f7043833a28959cad882a0a2884 model-00078-of-00162.safetensors
83
+ 3d7d835c3ec980321932997d0f13e09e6dce4366581781f57abe7f486de4724c model-00079-of-00162.safetensors
84
+ f9c14f3e927112dff49ec821a912a30fbcdda5c8f59a02445b15f1a9a948049f model-00080-of-00162.safetensors
85
+ bdb9827734211f3e21d469728fdbfc1a07ee4b853bdb07ed71a276b4ab6f3b56 model-00081-of-00162.safetensors
86
+ a0488e881985b942705561b4a3ccc43be61e8c4b7e1496a11c955a2de886f699 model-00082-of-00162.safetensors
87
+ bee4788577e142a785537ac156fa55df576e3b185b0d3209a16ff780b5018b40 model-00083-of-00162.safetensors
88
+ 006b23db27c1d51edb867f237b179c124fc2730d20e8016ef8e8507a8f70909b model-00084-of-00162.safetensors
89
+ c2021ec33c615a610a5269d1b58170ae2d0b92a71c8fc2e3c88ab0ffd69b6501 model-00085-of-00162.safetensors
90
+ a288fc73e64ad71080094feffa163770c69c7e688c6f2b719ff2c1f090acb38c model-00086-of-00162.safetensors
91
+ 4cdf8a79f610d63a56de2f6d28e6d99b2a41ec4e0c35b88b84e52a23db13b86e model-00087-of-00162.safetensors
92
+ 576eaf5c18fd579b2f698eea704982ee39ab9cbc2e622eb85b62b0430ab9f168 model-00088-of-00162.safetensors
93
+ 09d81765096aec67cf27082603d05e9825794621100a7d0add841f0fcd01252e model-00089-of-00162.safetensors
94
+ a298fe6606dd80aa30b97374e2d9caaba4a5d969837e2d1c8a8d62b224625116 model-00090-of-00162.safetensors
95
+ 4f62f72a35ddd85b496fecf750c76f6b8a3326bc8e5e1577b6525d01c8b4604c model-00091-of-00162.safetensors
96
+ 70d8a6707a6f63000ceb7ec742f6677bb442ff14d42e1f43da6c23e100f8e678 model-00092-of-00162.safetensors
97
+ 928a209606c16913c1d0a0ed163745e050fee6c82ca03cd93c4901074568c6d6 model-00093-of-00162.safetensors
98
+ 61aa87f0b15329b20429d476edbfc5582b6dbe39bc3bd008232a953fccc3ba91 model-00094-of-00162.safetensors
99
+ 9918962918997f1bc445e7f3c451b4b2b60329cc586bcb194d47b521de601ee5 model-00095-of-00162.safetensors
100
+ 5b05be10b0c45cfe910a13f63aff53b638a538b24b83b8edd3c653066d54ab9f model-00096-of-00162.safetensors
101
+ 7e8cca9aca5f90db748882cc92590d7f2f3fef8be4580c1605fd624f6f4848fe model-00097-of-00162.safetensors
102
+ fc5063dfb57fc7260bbf4623e1cd920948495dc957a78af9eed086429355992b model-00098-of-00162.safetensors
103
+ 83f746227391e256e1116e897b988e018c117cbf8ac52c963793d7448fc94aed model-00099-of-00162.safetensors
104
+ e5812c1b77a537c3bf7c5d798fe4c0791438b8bb321767249793bca0956c34f1 model-00100-of-00162.safetensors
105
+ d3d0c0578f59b2e1354bac1333d6c699df683dbede2cbb43e827e3de9fe2b4b5 model-00101-of-00162.safetensors
106
+ 98f4d0a9aed00e82a29aee965c6a3a45b43869035cdbab1603a549c54e4ffc8f model-00102-of-00162.safetensors
107
+ 1828269ab065765804bab5299a3b59a7d43eee93e750f81cec413d368d984267 model-00103-of-00162.safetensors
108
+ 9249a15faac58d4f6f5103262342358c1bc4297ad6895c763c677a4cc4792fd1 model-00104-of-00162.safetensors
109
+ c2dfd894e70870d70b9ca1377003070cfb640ef178d2ed4696e822a7ddd07b1d model-00105-of-00162.safetensors
110
+ cb00ed30c608035320c30a8238b4e6943d8fc7bc2abe30d402b31cd540236d57 model-00106-of-00162.safetensors
111
+ 9b3d2a381bee6203dc6f5eae65149f0bf487028aac2e79d251958681415d9575 model-00107-of-00162.safetensors
112
+ efd2ae02d74c113f67a2c1feba68ac165bd957dbb9492b5fcea95f2ea75c365f model-00108-of-00162.safetensors
113
+ facf62c5b5c380dc44d61f5b8aff9aa4e519b048efe42b3b947dbed7b319bc1a model-00109-of-00162.safetensors
114
+ 1431442e066fb232252555347c1c4cd14b109c4f2aef63a41ef8e2ba535d5977 model-00110-of-00162.safetensors
115
+ cf5155fa92289021ef8791aa4258bace6c2690236d971321a38dc723e2ef2854 model-00111-of-00162.safetensors
116
+ 3e3cc34ecb725d3e53aead557120d0ac88da94f79e96f12fde9dd5f27b34cc5a model-00112-of-00162.safetensors
117
+ 586d4c94f2f1980ea713d7988d866ca240665c44599d33e3b7569c8b58ad8745 model-00113-of-00162.safetensors
118
+ c908fe1fbdaa14b0acf5a4c390079ac05d86ff2956f8b82dd4654aeda138f16c model-00114-of-00162.safetensors
119
+ e7fc61963dd496d757f3cc0c1f7934461e6ad3c865820ceb8a29a6b784817851 model-00115-of-00162.safetensors
120
+ d2f9e1aecc2bca0c2b678da1b720ab4253c35ecaa1cbc42983e8ff9800ef37f4 model-00116-of-00162.safetensors
121
+ b06fbff515f9c6ff6a4a714a52cc6976a82b93a04a98f650a478df5504465f34 model-00117-of-00162.safetensors
122
+ 1dd357f1a647ddf5b306d955dceec031da73787ba08014a7577d5c1dd86fad1c model-00118-of-00162.safetensors
123
+ dccc9decdb71aca8355ee516180a8465780f6211ddefa5908b40ac69c12a0357 model-00119-of-00162.safetensors
124
+ 1b501bd5f034c9af1cbff096e0036e246d66c42520a509aed6cf792339b2495c model-00120-of-00162.safetensors
125
+ ecd0f3322aa147b7936e34ab2dde44698716b3186dd4818a7be0da80d18e6488 model-00121-of-00162.safetensors
126
+ 943e5693f2f476656ad70134df1173a2350633eb2a21a754d5cd905c5bf57e94 model-00122-of-00162.safetensors
127
+ 4612b734b7bf4e5d2a94ef1d93f854356965320ec3445e4a189f46a3cdb7faf1 model-00123-of-00162.safetensors
128
+ 04604658ca7d8160a6ae846f147e314a9e29d49895e447d6987dc060e33df2a1 model-00124-of-00162.safetensors
129
+ 775c87f6b1862c2e88a98be2de75b76e3bdb3980b47313eaa0361808501a3508 model-00125-of-00162.safetensors
130
+ 0e3f1740885bed2586e2660b1cfe408cf15df3989eccc20ab77fed5831945226 model-00126-of-00162.safetensors
131
+ 5656fe83a4813980005e996defc2498f2baeb4161299067cad0bf1b94b2eea58 model-00127-of-00162.safetensors
132
+ 63fe5db5f2cd78f4811749785cd69ec0c936263598912a94a27c9787be39ca7d model-00128-of-00162.safetensors
133
+ 663d345e2bf95475752414dfe3338c18ca47f562ccdd09d6c38fb065fe50a669 model-00129-of-00162.safetensors
134
+ d1bcea5450d23604e2413e0bc1293f24d3939fa0d03d13c3595542f62a55b5fa model-00130-of-00162.safetensors
135
+ aab17b0b79d17c4d80cd6f61648d6d802ba7627ea5e0d40ce35a83363dd7fc4b model-00131-of-00162.safetensors
136
+ c3d54f3f1c447518abb33d91fd351ec86a2b4f370a18a1c1f3cd94b90a3a4781 model-00132-of-00162.safetensors
137
+ 2054e7e3428652f6cf0145b380d45d8df4b6bd1daa0082f93a428f14588e8864 model-00133-of-00162.safetensors
138
+ 71299301b48ac61a53b09624218365a57219aa62f854e1dc318e418c08f77597 model-00134-of-00162.safetensors
139
+ 55d0abd7f13059f8af7584c634530ec567ee6a22de0c8704544c7896af8c32c0 model-00135-of-00162.safetensors
140
+ 928fce84e7c5234eeaae5ee9b5a8bc8d32dfaed8ac3fbd2a9c279f1a76d76bae model-00136-of-00162.safetensors
141
+ 76fd468f3ed69de9639a38f8e6128119457209e5cf68101676872294e13425aa model-00137-of-00162.safetensors
142
+ a2a6484a6eed83dded51aaaaa0d47c4959027b41dbc31c80cd28e68fb7b085f1 model-00138-of-00162.safetensors
143
+ fd73326d76b3755b23ce3280a1498471c3c0feb59eb270e2b66413c6e10d1d74 model-00139-of-00162.safetensors
144
+ d243874f8aad4e682b0fe741373df42e3d5744dd1d9237de26bb8697913dde59 model-00140-of-00162.safetensors
145
+ 209cfc9251cd001a62f1ae15f92f073e40c7651febddf2306ff2133003eecfd1 model-00141-of-00162.safetensors
146
+ 7671eae4c57760631be6e5a51971fe58a0aea036467b62d58505ec24b9278d0f model-00142-of-00162.safetensors
147
+ 29ab2e273f72400c8bbddc593a18b855325d8c207ba14ea0c49e0b89257dc9c1 model-00143-of-00162.safetensors
148
+ ce8117b5c680ab1223d450aaab61abe1c3cfa39744f6776a9f7ce91dc885628c model-00144-of-00162.safetensors
149
+ 2c24fd7f9323c533b89816e78d58c81c6cea751081db833beed101fbd8cf17fb model-00145-of-00162.safetensors
150
+ 673e4451c8d7b66cbb65a88b7367df182886e3e42070963b7812d30023a1c767 model-00146-of-00162.safetensors
151
+ 9d7914a5b30d5459f2c349680dea4dc2652985b958852cb422256d6dc48f4d13 model-00147-of-00162.safetensors
152
+ 95939bb6c264344cfb540c7f78829a7f6d0635804c21ea33e98d554d99802ab1 model-00148-of-00162.safetensors
153
+ cde9742fa621eb926907b1910f8f32d2bb0dce4c3f2f4d60e794969795de2297 model-00149-of-00162.safetensors
154
+ d5a5577f7c8c009b9564f29d6894104b2e5f6e5b0af74b05b4679ffc8850540c model-00150-of-00162.safetensors
155
+ 91cb7249bac44c8e8f5b39fc57a739dd3ec2d98dd925d931f769a213d7644af8 model-00151-of-00162.safetensors
156
+ a3640d66b5e2a06ae2b5b248fef79a95019a7f7c32bb18c144da040caf23d435 model-00152-of-00162.safetensors
157
+ 7ea1e2c61d6e7708a51759991448355a48f4459ca4e59b50a35c50af28a55187 model-00153-of-00162.safetensors
158
+ f546cebf96f091599dd8bcefe842709ce845e7f597bafb3d11b492b5d3fea3cc model-00154-of-00162.safetensors
159
+ 699ca083bf794f871271ccadd31e3b0ac4428f86822092ad6fa8ff789256eb3c model-00155-of-00162.safetensors
160
+ 1f9e63f4539817b264b888d8753b3644c9ba1799d1a5e4d8845e78a68cf87d54 model-00156-of-00162.safetensors
161
+ 46b87e3bfdc1ccdb570f5b0379f133e4f01d2315760143a6f4d426bce48de975 model-00157-of-00162.safetensors
162
+ bc635e7f76da8fd5326b6bfe79217ab58ca43ca72e52ad5d15a6b526f698cd2f model-00158-of-00162.safetensors
163
+ 4c27fe83e62e9cf179081ff589877d229a1c5916a7817d266b80be1388fdf875 model-00159-of-00162.safetensors
164
+ f7093275e89922a19070b8ce8e157044589206066327aac8e75c7bddcf40f44c model-00160-of-00162.safetensors
165
+ fcc593f9740dcf8c1198e73b92dd1fbaf97924a5e5eec6ff12dbe7f6eb7cf790 model-00161-of-00162.safetensors
166
+ 230a9e3582b5e31b24bcd60c1e3f47bfc151622a1fd5e8c9bd480eebc29e7e55 model-00162-of-00162.safetensors
measurement.json ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors.index.json ADDED
@@ -0,0 +1,730 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 282214825984
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00161-of-00162.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00162.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00003-of-00162.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00003-of-00162.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00162.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00003-of-00162.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00003-of-00162.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00162.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00162.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00162.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00162.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00005-of-00162.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00005-of-00162.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00004-of-00162.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00005-of-00162.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00005-of-00162.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00004-of-00162.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00004-of-00162.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00004-of-00162.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00004-of-00162.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00023-of-00162.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00023-of-00162.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00022-of-00162.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00023-of-00162.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00023-of-00162.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00022-of-00162.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00022-of-00162.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00022-of-00162.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00022-of-00162.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00025-of-00162.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00025-of-00162.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00024-of-00162.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00025-of-00162.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00025-of-00162.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00024-of-00162.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00024-of-00162.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00024-of-00162.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00024-of-00162.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00027-of-00162.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00027-of-00162.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00026-of-00162.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00027-of-00162.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00027-of-00162.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00026-of-00162.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00026-of-00162.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00026-of-00162.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00026-of-00162.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00029-of-00162.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00029-of-00162.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00028-of-00162.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00029-of-00162.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00029-of-00162.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00028-of-00162.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00028-of-00162.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00028-of-00162.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00028-of-00162.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00031-of-00162.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00031-of-00162.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00030-of-00162.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00031-of-00162.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00031-of-00162.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00030-of-00162.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00030-of-00162.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00030-of-00162.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00030-of-00162.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00033-of-00162.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00033-of-00162.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00032-of-00162.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00033-of-00162.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00033-of-00162.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00032-of-00162.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00032-of-00162.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00032-of-00162.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00032-of-00162.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00035-of-00162.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00035-of-00162.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00034-of-00162.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00035-of-00162.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00035-of-00162.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00034-of-00162.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00034-of-00162.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00034-of-00162.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00034-of-00162.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00037-of-00162.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00037-of-00162.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00036-of-00162.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00037-of-00162.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00037-of-00162.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00036-of-00162.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00036-of-00162.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00036-of-00162.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00036-of-00162.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00039-of-00162.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00039-of-00162.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00038-of-00162.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00039-of-00162.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00039-of-00162.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00038-of-00162.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00038-of-00162.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00038-of-00162.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00038-of-00162.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00041-of-00162.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00041-of-00162.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00040-of-00162.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00041-of-00162.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00041-of-00162.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00040-of-00162.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00040-of-00162.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00040-of-00162.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00040-of-00162.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00007-of-00162.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00007-of-00162.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00006-of-00162.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00007-of-00162.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00007-of-00162.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00006-of-00162.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00006-of-00162.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00006-of-00162.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00006-of-00162.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00043-of-00162.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00043-of-00162.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00042-of-00162.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00043-of-00162.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00043-of-00162.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00042-of-00162.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00042-of-00162.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00042-of-00162.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00042-of-00162.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00045-of-00162.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00045-of-00162.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00044-of-00162.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00045-of-00162.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00045-of-00162.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00044-of-00162.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00044-of-00162.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00044-of-00162.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00044-of-00162.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00047-of-00162.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00047-of-00162.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00046-of-00162.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00047-of-00162.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00047-of-00162.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00046-of-00162.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00046-of-00162.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00046-of-00162.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00046-of-00162.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00049-of-00162.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00049-of-00162.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00048-of-00162.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00049-of-00162.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00049-of-00162.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00048-of-00162.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00048-of-00162.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00048-of-00162.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00048-of-00162.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00051-of-00162.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00051-of-00162.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00050-of-00162.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00051-of-00162.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00051-of-00162.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00050-of-00162.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00050-of-00162.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00050-of-00162.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00050-of-00162.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00053-of-00162.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00053-of-00162.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00052-of-00162.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00053-of-00162.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00053-of-00162.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00052-of-00162.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00052-of-00162.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00052-of-00162.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00052-of-00162.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00055-of-00162.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00055-of-00162.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00054-of-00162.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00055-of-00162.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00055-of-00162.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00054-of-00162.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00054-of-00162.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00054-of-00162.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00054-of-00162.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00057-of-00162.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00057-of-00162.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00056-of-00162.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00057-of-00162.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00057-of-00162.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00056-of-00162.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00056-of-00162.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00056-of-00162.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00056-of-00162.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00059-of-00162.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00059-of-00162.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00058-of-00162.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00059-of-00162.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00059-of-00162.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00058-of-00162.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00058-of-00162.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00058-of-00162.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00058-of-00162.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00061-of-00162.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00061-of-00162.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00060-of-00162.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00061-of-00162.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00061-of-00162.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00060-of-00162.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00060-of-00162.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00060-of-00162.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00060-of-00162.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00009-of-00162.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00009-of-00162.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00008-of-00162.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00009-of-00162.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00009-of-00162.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00008-of-00162.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00008-of-00162.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00008-of-00162.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00008-of-00162.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00063-of-00162.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00063-of-00162.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00062-of-00162.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00063-of-00162.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00063-of-00162.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00062-of-00162.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00062-of-00162.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00062-of-00162.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00062-of-00162.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00065-of-00162.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00065-of-00162.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00064-of-00162.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00065-of-00162.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00065-of-00162.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00064-of-00162.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00064-of-00162.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00064-of-00162.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00064-of-00162.safetensors",
242
+ "model.layers.32.input_layernorm.weight": "model-00067-of-00162.safetensors",
243
+ "model.layers.32.mlp.down_proj.weight": "model-00067-of-00162.safetensors",
244
+ "model.layers.32.mlp.gate_proj.weight": "model-00066-of-00162.safetensors",
245
+ "model.layers.32.mlp.up_proj.weight": "model-00067-of-00162.safetensors",
246
+ "model.layers.32.post_attention_layernorm.weight": "model-00067-of-00162.safetensors",
247
+ "model.layers.32.self_attn.k_proj.weight": "model-00066-of-00162.safetensors",
248
+ "model.layers.32.self_attn.o_proj.weight": "model-00066-of-00162.safetensors",
249
+ "model.layers.32.self_attn.q_proj.weight": "model-00066-of-00162.safetensors",
250
+ "model.layers.32.self_attn.v_proj.weight": "model-00066-of-00162.safetensors",
251
+ "model.layers.33.input_layernorm.weight": "model-00069-of-00162.safetensors",
252
+ "model.layers.33.mlp.down_proj.weight": "model-00069-of-00162.safetensors",
253
+ "model.layers.33.mlp.gate_proj.weight": "model-00068-of-00162.safetensors",
254
+ "model.layers.33.mlp.up_proj.weight": "model-00069-of-00162.safetensors",
255
+ "model.layers.33.post_attention_layernorm.weight": "model-00069-of-00162.safetensors",
256
+ "model.layers.33.self_attn.k_proj.weight": "model-00068-of-00162.safetensors",
257
+ "model.layers.33.self_attn.o_proj.weight": "model-00068-of-00162.safetensors",
258
+ "model.layers.33.self_attn.q_proj.weight": "model-00068-of-00162.safetensors",
259
+ "model.layers.33.self_attn.v_proj.weight": "model-00068-of-00162.safetensors",
260
+ "model.layers.34.input_layernorm.weight": "model-00071-of-00162.safetensors",
261
+ "model.layers.34.mlp.down_proj.weight": "model-00071-of-00162.safetensors",
262
+ "model.layers.34.mlp.gate_proj.weight": "model-00070-of-00162.safetensors",
263
+ "model.layers.34.mlp.up_proj.weight": "model-00071-of-00162.safetensors",
264
+ "model.layers.34.post_attention_layernorm.weight": "model-00071-of-00162.safetensors",
265
+ "model.layers.34.self_attn.k_proj.weight": "model-00070-of-00162.safetensors",
266
+ "model.layers.34.self_attn.o_proj.weight": "model-00070-of-00162.safetensors",
267
+ "model.layers.34.self_attn.q_proj.weight": "model-00070-of-00162.safetensors",
268
+ "model.layers.34.self_attn.v_proj.weight": "model-00070-of-00162.safetensors",
269
+ "model.layers.35.input_layernorm.weight": "model-00073-of-00162.safetensors",
270
+ "model.layers.35.mlp.down_proj.weight": "model-00073-of-00162.safetensors",
271
+ "model.layers.35.mlp.gate_proj.weight": "model-00072-of-00162.safetensors",
272
+ "model.layers.35.mlp.up_proj.weight": "model-00073-of-00162.safetensors",
273
+ "model.layers.35.post_attention_layernorm.weight": "model-00073-of-00162.safetensors",
274
+ "model.layers.35.self_attn.k_proj.weight": "model-00072-of-00162.safetensors",
275
+ "model.layers.35.self_attn.o_proj.weight": "model-00072-of-00162.safetensors",
276
+ "model.layers.35.self_attn.q_proj.weight": "model-00072-of-00162.safetensors",
277
+ "model.layers.35.self_attn.v_proj.weight": "model-00072-of-00162.safetensors",
278
+ "model.layers.36.input_layernorm.weight": "model-00075-of-00162.safetensors",
279
+ "model.layers.36.mlp.down_proj.weight": "model-00075-of-00162.safetensors",
280
+ "model.layers.36.mlp.gate_proj.weight": "model-00074-of-00162.safetensors",
281
+ "model.layers.36.mlp.up_proj.weight": "model-00075-of-00162.safetensors",
282
+ "model.layers.36.post_attention_layernorm.weight": "model-00075-of-00162.safetensors",
283
+ "model.layers.36.self_attn.k_proj.weight": "model-00074-of-00162.safetensors",
284
+ "model.layers.36.self_attn.o_proj.weight": "model-00074-of-00162.safetensors",
285
+ "model.layers.36.self_attn.q_proj.weight": "model-00074-of-00162.safetensors",
286
+ "model.layers.36.self_attn.v_proj.weight": "model-00074-of-00162.safetensors",
287
+ "model.layers.37.input_layernorm.weight": "model-00077-of-00162.safetensors",
288
+ "model.layers.37.mlp.down_proj.weight": "model-00077-of-00162.safetensors",
289
+ "model.layers.37.mlp.gate_proj.weight": "model-00076-of-00162.safetensors",
290
+ "model.layers.37.mlp.up_proj.weight": "model-00077-of-00162.safetensors",
291
+ "model.layers.37.post_attention_layernorm.weight": "model-00077-of-00162.safetensors",
292
+ "model.layers.37.self_attn.k_proj.weight": "model-00076-of-00162.safetensors",
293
+ "model.layers.37.self_attn.o_proj.weight": "model-00076-of-00162.safetensors",
294
+ "model.layers.37.self_attn.q_proj.weight": "model-00076-of-00162.safetensors",
295
+ "model.layers.37.self_attn.v_proj.weight": "model-00076-of-00162.safetensors",
296
+ "model.layers.38.input_layernorm.weight": "model-00079-of-00162.safetensors",
297
+ "model.layers.38.mlp.down_proj.weight": "model-00079-of-00162.safetensors",
298
+ "model.layers.38.mlp.gate_proj.weight": "model-00078-of-00162.safetensors",
299
+ "model.layers.38.mlp.up_proj.weight": "model-00079-of-00162.safetensors",
300
+ "model.layers.38.post_attention_layernorm.weight": "model-00079-of-00162.safetensors",
301
+ "model.layers.38.self_attn.k_proj.weight": "model-00078-of-00162.safetensors",
302
+ "model.layers.38.self_attn.o_proj.weight": "model-00078-of-00162.safetensors",
303
+ "model.layers.38.self_attn.q_proj.weight": "model-00078-of-00162.safetensors",
304
+ "model.layers.38.self_attn.v_proj.weight": "model-00078-of-00162.safetensors",
305
+ "model.layers.39.input_layernorm.weight": "model-00081-of-00162.safetensors",
306
+ "model.layers.39.mlp.down_proj.weight": "model-00081-of-00162.safetensors",
307
+ "model.layers.39.mlp.gate_proj.weight": "model-00080-of-00162.safetensors",
308
+ "model.layers.39.mlp.up_proj.weight": "model-00081-of-00162.safetensors",
309
+ "model.layers.39.post_attention_layernorm.weight": "model-00081-of-00162.safetensors",
310
+ "model.layers.39.self_attn.k_proj.weight": "model-00080-of-00162.safetensors",
311
+ "model.layers.39.self_attn.o_proj.weight": "model-00080-of-00162.safetensors",
312
+ "model.layers.39.self_attn.q_proj.weight": "model-00080-of-00162.safetensors",
313
+ "model.layers.39.self_attn.v_proj.weight": "model-00080-of-00162.safetensors",
314
+ "model.layers.4.input_layernorm.weight": "model-00011-of-00162.safetensors",
315
+ "model.layers.4.mlp.down_proj.weight": "model-00011-of-00162.safetensors",
316
+ "model.layers.4.mlp.gate_proj.weight": "model-00010-of-00162.safetensors",
317
+ "model.layers.4.mlp.up_proj.weight": "model-00011-of-00162.safetensors",
318
+ "model.layers.4.post_attention_layernorm.weight": "model-00011-of-00162.safetensors",
319
+ "model.layers.4.self_attn.k_proj.weight": "model-00010-of-00162.safetensors",
320
+ "model.layers.4.self_attn.o_proj.weight": "model-00010-of-00162.safetensors",
321
+ "model.layers.4.self_attn.q_proj.weight": "model-00010-of-00162.safetensors",
322
+ "model.layers.4.self_attn.v_proj.weight": "model-00010-of-00162.safetensors",
323
+ "model.layers.40.input_layernorm.weight": "model-00083-of-00162.safetensors",
324
+ "model.layers.40.mlp.down_proj.weight": "model-00083-of-00162.safetensors",
325
+ "model.layers.40.mlp.gate_proj.weight": "model-00082-of-00162.safetensors",
326
+ "model.layers.40.mlp.up_proj.weight": "model-00083-of-00162.safetensors",
327
+ "model.layers.40.post_attention_layernorm.weight": "model-00083-of-00162.safetensors",
328
+ "model.layers.40.self_attn.k_proj.weight": "model-00082-of-00162.safetensors",
329
+ "model.layers.40.self_attn.o_proj.weight": "model-00082-of-00162.safetensors",
330
+ "model.layers.40.self_attn.q_proj.weight": "model-00082-of-00162.safetensors",
331
+ "model.layers.40.self_attn.v_proj.weight": "model-00082-of-00162.safetensors",
332
+ "model.layers.41.input_layernorm.weight": "model-00085-of-00162.safetensors",
333
+ "model.layers.41.mlp.down_proj.weight": "model-00085-of-00162.safetensors",
334
+ "model.layers.41.mlp.gate_proj.weight": "model-00084-of-00162.safetensors",
335
+ "model.layers.41.mlp.up_proj.weight": "model-00085-of-00162.safetensors",
336
+ "model.layers.41.post_attention_layernorm.weight": "model-00085-of-00162.safetensors",
337
+ "model.layers.41.self_attn.k_proj.weight": "model-00084-of-00162.safetensors",
338
+ "model.layers.41.self_attn.o_proj.weight": "model-00084-of-00162.safetensors",
339
+ "model.layers.41.self_attn.q_proj.weight": "model-00084-of-00162.safetensors",
340
+ "model.layers.41.self_attn.v_proj.weight": "model-00084-of-00162.safetensors",
341
+ "model.layers.42.input_layernorm.weight": "model-00087-of-00162.safetensors",
342
+ "model.layers.42.mlp.down_proj.weight": "model-00087-of-00162.safetensors",
343
+ "model.layers.42.mlp.gate_proj.weight": "model-00086-of-00162.safetensors",
344
+ "model.layers.42.mlp.up_proj.weight": "model-00087-of-00162.safetensors",
345
+ "model.layers.42.post_attention_layernorm.weight": "model-00087-of-00162.safetensors",
346
+ "model.layers.42.self_attn.k_proj.weight": "model-00086-of-00162.safetensors",
347
+ "model.layers.42.self_attn.o_proj.weight": "model-00086-of-00162.safetensors",
348
+ "model.layers.42.self_attn.q_proj.weight": "model-00086-of-00162.safetensors",
349
+ "model.layers.42.self_attn.v_proj.weight": "model-00086-of-00162.safetensors",
350
+ "model.layers.43.input_layernorm.weight": "model-00089-of-00162.safetensors",
351
+ "model.layers.43.mlp.down_proj.weight": "model-00089-of-00162.safetensors",
352
+ "model.layers.43.mlp.gate_proj.weight": "model-00088-of-00162.safetensors",
353
+ "model.layers.43.mlp.up_proj.weight": "model-00089-of-00162.safetensors",
354
+ "model.layers.43.post_attention_layernorm.weight": "model-00089-of-00162.safetensors",
355
+ "model.layers.43.self_attn.k_proj.weight": "model-00088-of-00162.safetensors",
356
+ "model.layers.43.self_attn.o_proj.weight": "model-00088-of-00162.safetensors",
357
+ "model.layers.43.self_attn.q_proj.weight": "model-00088-of-00162.safetensors",
358
+ "model.layers.43.self_attn.v_proj.weight": "model-00088-of-00162.safetensors",
359
+ "model.layers.44.input_layernorm.weight": "model-00091-of-00162.safetensors",
360
+ "model.layers.44.mlp.down_proj.weight": "model-00091-of-00162.safetensors",
361
+ "model.layers.44.mlp.gate_proj.weight": "model-00090-of-00162.safetensors",
362
+ "model.layers.44.mlp.up_proj.weight": "model-00091-of-00162.safetensors",
363
+ "model.layers.44.post_attention_layernorm.weight": "model-00091-of-00162.safetensors",
364
+ "model.layers.44.self_attn.k_proj.weight": "model-00090-of-00162.safetensors",
365
+ "model.layers.44.self_attn.o_proj.weight": "model-00090-of-00162.safetensors",
366
+ "model.layers.44.self_attn.q_proj.weight": "model-00090-of-00162.safetensors",
367
+ "model.layers.44.self_attn.v_proj.weight": "model-00090-of-00162.safetensors",
368
+ "model.layers.45.input_layernorm.weight": "model-00093-of-00162.safetensors",
369
+ "model.layers.45.mlp.down_proj.weight": "model-00093-of-00162.safetensors",
370
+ "model.layers.45.mlp.gate_proj.weight": "model-00092-of-00162.safetensors",
371
+ "model.layers.45.mlp.up_proj.weight": "model-00093-of-00162.safetensors",
372
+ "model.layers.45.post_attention_layernorm.weight": "model-00093-of-00162.safetensors",
373
+ "model.layers.45.self_attn.k_proj.weight": "model-00092-of-00162.safetensors",
374
+ "model.layers.45.self_attn.o_proj.weight": "model-00092-of-00162.safetensors",
375
+ "model.layers.45.self_attn.q_proj.weight": "model-00092-of-00162.safetensors",
376
+ "model.layers.45.self_attn.v_proj.weight": "model-00092-of-00162.safetensors",
377
+ "model.layers.46.input_layernorm.weight": "model-00095-of-00162.safetensors",
378
+ "model.layers.46.mlp.down_proj.weight": "model-00095-of-00162.safetensors",
379
+ "model.layers.46.mlp.gate_proj.weight": "model-00094-of-00162.safetensors",
380
+ "model.layers.46.mlp.up_proj.weight": "model-00095-of-00162.safetensors",
381
+ "model.layers.46.post_attention_layernorm.weight": "model-00095-of-00162.safetensors",
382
+ "model.layers.46.self_attn.k_proj.weight": "model-00094-of-00162.safetensors",
383
+ "model.layers.46.self_attn.o_proj.weight": "model-00094-of-00162.safetensors",
384
+ "model.layers.46.self_attn.q_proj.weight": "model-00094-of-00162.safetensors",
385
+ "model.layers.46.self_attn.v_proj.weight": "model-00094-of-00162.safetensors",
386
+ "model.layers.47.input_layernorm.weight": "model-00097-of-00162.safetensors",
387
+ "model.layers.47.mlp.down_proj.weight": "model-00097-of-00162.safetensors",
388
+ "model.layers.47.mlp.gate_proj.weight": "model-00096-of-00162.safetensors",
389
+ "model.layers.47.mlp.up_proj.weight": "model-00097-of-00162.safetensors",
390
+ "model.layers.47.post_attention_layernorm.weight": "model-00097-of-00162.safetensors",
391
+ "model.layers.47.self_attn.k_proj.weight": "model-00096-of-00162.safetensors",
392
+ "model.layers.47.self_attn.o_proj.weight": "model-00096-of-00162.safetensors",
393
+ "model.layers.47.self_attn.q_proj.weight": "model-00096-of-00162.safetensors",
394
+ "model.layers.47.self_attn.v_proj.weight": "model-00096-of-00162.safetensors",
395
+ "model.layers.48.input_layernorm.weight": "model-00099-of-00162.safetensors",
396
+ "model.layers.48.mlp.down_proj.weight": "model-00099-of-00162.safetensors",
397
+ "model.layers.48.mlp.gate_proj.weight": "model-00098-of-00162.safetensors",
398
+ "model.layers.48.mlp.up_proj.weight": "model-00099-of-00162.safetensors",
399
+ "model.layers.48.post_attention_layernorm.weight": "model-00099-of-00162.safetensors",
400
+ "model.layers.48.self_attn.k_proj.weight": "model-00098-of-00162.safetensors",
401
+ "model.layers.48.self_attn.o_proj.weight": "model-00098-of-00162.safetensors",
402
+ "model.layers.48.self_attn.q_proj.weight": "model-00098-of-00162.safetensors",
403
+ "model.layers.48.self_attn.v_proj.weight": "model-00098-of-00162.safetensors",
404
+ "model.layers.49.input_layernorm.weight": "model-00101-of-00162.safetensors",
405
+ "model.layers.49.mlp.down_proj.weight": "model-00101-of-00162.safetensors",
406
+ "model.layers.49.mlp.gate_proj.weight": "model-00100-of-00162.safetensors",
407
+ "model.layers.49.mlp.up_proj.weight": "model-00101-of-00162.safetensors",
408
+ "model.layers.49.post_attention_layernorm.weight": "model-00101-of-00162.safetensors",
409
+ "model.layers.49.self_attn.k_proj.weight": "model-00100-of-00162.safetensors",
410
+ "model.layers.49.self_attn.o_proj.weight": "model-00100-of-00162.safetensors",
411
+ "model.layers.49.self_attn.q_proj.weight": "model-00100-of-00162.safetensors",
412
+ "model.layers.49.self_attn.v_proj.weight": "model-00100-of-00162.safetensors",
413
+ "model.layers.5.input_layernorm.weight": "model-00013-of-00162.safetensors",
414
+ "model.layers.5.mlp.down_proj.weight": "model-00013-of-00162.safetensors",
415
+ "model.layers.5.mlp.gate_proj.weight": "model-00012-of-00162.safetensors",
416
+ "model.layers.5.mlp.up_proj.weight": "model-00013-of-00162.safetensors",
417
+ "model.layers.5.post_attention_layernorm.weight": "model-00013-of-00162.safetensors",
418
+ "model.layers.5.self_attn.k_proj.weight": "model-00012-of-00162.safetensors",
419
+ "model.layers.5.self_attn.o_proj.weight": "model-00012-of-00162.safetensors",
420
+ "model.layers.5.self_attn.q_proj.weight": "model-00012-of-00162.safetensors",
421
+ "model.layers.5.self_attn.v_proj.weight": "model-00012-of-00162.safetensors",
422
+ "model.layers.50.input_layernorm.weight": "model-00103-of-00162.safetensors",
423
+ "model.layers.50.mlp.down_proj.weight": "model-00103-of-00162.safetensors",
424
+ "model.layers.50.mlp.gate_proj.weight": "model-00102-of-00162.safetensors",
425
+ "model.layers.50.mlp.up_proj.weight": "model-00103-of-00162.safetensors",
426
+ "model.layers.50.post_attention_layernorm.weight": "model-00103-of-00162.safetensors",
427
+ "model.layers.50.self_attn.k_proj.weight": "model-00102-of-00162.safetensors",
428
+ "model.layers.50.self_attn.o_proj.weight": "model-00102-of-00162.safetensors",
429
+ "model.layers.50.self_attn.q_proj.weight": "model-00102-of-00162.safetensors",
430
+ "model.layers.50.self_attn.v_proj.weight": "model-00102-of-00162.safetensors",
431
+ "model.layers.51.input_layernorm.weight": "model-00105-of-00162.safetensors",
432
+ "model.layers.51.mlp.down_proj.weight": "model-00105-of-00162.safetensors",
433
+ "model.layers.51.mlp.gate_proj.weight": "model-00104-of-00162.safetensors",
434
+ "model.layers.51.mlp.up_proj.weight": "model-00105-of-00162.safetensors",
435
+ "model.layers.51.post_attention_layernorm.weight": "model-00105-of-00162.safetensors",
436
+ "model.layers.51.self_attn.k_proj.weight": "model-00104-of-00162.safetensors",
437
+ "model.layers.51.self_attn.o_proj.weight": "model-00104-of-00162.safetensors",
438
+ "model.layers.51.self_attn.q_proj.weight": "model-00104-of-00162.safetensors",
439
+ "model.layers.51.self_attn.v_proj.weight": "model-00104-of-00162.safetensors",
440
+ "model.layers.52.input_layernorm.weight": "model-00107-of-00162.safetensors",
441
+ "model.layers.52.mlp.down_proj.weight": "model-00107-of-00162.safetensors",
442
+ "model.layers.52.mlp.gate_proj.weight": "model-00106-of-00162.safetensors",
443
+ "model.layers.52.mlp.up_proj.weight": "model-00107-of-00162.safetensors",
444
+ "model.layers.52.post_attention_layernorm.weight": "model-00107-of-00162.safetensors",
445
+ "model.layers.52.self_attn.k_proj.weight": "model-00106-of-00162.safetensors",
446
+ "model.layers.52.self_attn.o_proj.weight": "model-00106-of-00162.safetensors",
447
+ "model.layers.52.self_attn.q_proj.weight": "model-00106-of-00162.safetensors",
448
+ "model.layers.52.self_attn.v_proj.weight": "model-00106-of-00162.safetensors",
449
+ "model.layers.53.input_layernorm.weight": "model-00109-of-00162.safetensors",
450
+ "model.layers.53.mlp.down_proj.weight": "model-00109-of-00162.safetensors",
451
+ "model.layers.53.mlp.gate_proj.weight": "model-00108-of-00162.safetensors",
452
+ "model.layers.53.mlp.up_proj.weight": "model-00109-of-00162.safetensors",
453
+ "model.layers.53.post_attention_layernorm.weight": "model-00109-of-00162.safetensors",
454
+ "model.layers.53.self_attn.k_proj.weight": "model-00108-of-00162.safetensors",
455
+ "model.layers.53.self_attn.o_proj.weight": "model-00108-of-00162.safetensors",
456
+ "model.layers.53.self_attn.q_proj.weight": "model-00108-of-00162.safetensors",
457
+ "model.layers.53.self_attn.v_proj.weight": "model-00108-of-00162.safetensors",
458
+ "model.layers.54.input_layernorm.weight": "model-00111-of-00162.safetensors",
459
+ "model.layers.54.mlp.down_proj.weight": "model-00111-of-00162.safetensors",
460
+ "model.layers.54.mlp.gate_proj.weight": "model-00110-of-00162.safetensors",
461
+ "model.layers.54.mlp.up_proj.weight": "model-00111-of-00162.safetensors",
462
+ "model.layers.54.post_attention_layernorm.weight": "model-00111-of-00162.safetensors",
463
+ "model.layers.54.self_attn.k_proj.weight": "model-00110-of-00162.safetensors",
464
+ "model.layers.54.self_attn.o_proj.weight": "model-00110-of-00162.safetensors",
465
+ "model.layers.54.self_attn.q_proj.weight": "model-00110-of-00162.safetensors",
466
+ "model.layers.54.self_attn.v_proj.weight": "model-00110-of-00162.safetensors",
467
+ "model.layers.55.input_layernorm.weight": "model-00113-of-00162.safetensors",
468
+ "model.layers.55.mlp.down_proj.weight": "model-00113-of-00162.safetensors",
469
+ "model.layers.55.mlp.gate_proj.weight": "model-00112-of-00162.safetensors",
470
+ "model.layers.55.mlp.up_proj.weight": "model-00113-of-00162.safetensors",
471
+ "model.layers.55.post_attention_layernorm.weight": "model-00113-of-00162.safetensors",
472
+ "model.layers.55.self_attn.k_proj.weight": "model-00112-of-00162.safetensors",
473
+ "model.layers.55.self_attn.o_proj.weight": "model-00112-of-00162.safetensors",
474
+ "model.layers.55.self_attn.q_proj.weight": "model-00112-of-00162.safetensors",
475
+ "model.layers.55.self_attn.v_proj.weight": "model-00112-of-00162.safetensors",
476
+ "model.layers.56.input_layernorm.weight": "model-00115-of-00162.safetensors",
477
+ "model.layers.56.mlp.down_proj.weight": "model-00115-of-00162.safetensors",
478
+ "model.layers.56.mlp.gate_proj.weight": "model-00114-of-00162.safetensors",
479
+ "model.layers.56.mlp.up_proj.weight": "model-00115-of-00162.safetensors",
480
+ "model.layers.56.post_attention_layernorm.weight": "model-00115-of-00162.safetensors",
481
+ "model.layers.56.self_attn.k_proj.weight": "model-00114-of-00162.safetensors",
482
+ "model.layers.56.self_attn.o_proj.weight": "model-00114-of-00162.safetensors",
483
+ "model.layers.56.self_attn.q_proj.weight": "model-00114-of-00162.safetensors",
484
+ "model.layers.56.self_attn.v_proj.weight": "model-00114-of-00162.safetensors",
485
+ "model.layers.57.input_layernorm.weight": "model-00117-of-00162.safetensors",
486
+ "model.layers.57.mlp.down_proj.weight": "model-00117-of-00162.safetensors",
487
+ "model.layers.57.mlp.gate_proj.weight": "model-00116-of-00162.safetensors",
488
+ "model.layers.57.mlp.up_proj.weight": "model-00117-of-00162.safetensors",
489
+ "model.layers.57.post_attention_layernorm.weight": "model-00117-of-00162.safetensors",
490
+ "model.layers.57.self_attn.k_proj.weight": "model-00116-of-00162.safetensors",
491
+ "model.layers.57.self_attn.o_proj.weight": "model-00116-of-00162.safetensors",
492
+ "model.layers.57.self_attn.q_proj.weight": "model-00116-of-00162.safetensors",
493
+ "model.layers.57.self_attn.v_proj.weight": "model-00116-of-00162.safetensors",
494
+ "model.layers.58.input_layernorm.weight": "model-00119-of-00162.safetensors",
495
+ "model.layers.58.mlp.down_proj.weight": "model-00119-of-00162.safetensors",
496
+ "model.layers.58.mlp.gate_proj.weight": "model-00118-of-00162.safetensors",
497
+ "model.layers.58.mlp.up_proj.weight": "model-00119-of-00162.safetensors",
498
+ "model.layers.58.post_attention_layernorm.weight": "model-00119-of-00162.safetensors",
499
+ "model.layers.58.self_attn.k_proj.weight": "model-00118-of-00162.safetensors",
500
+ "model.layers.58.self_attn.o_proj.weight": "model-00118-of-00162.safetensors",
501
+ "model.layers.58.self_attn.q_proj.weight": "model-00118-of-00162.safetensors",
502
+ "model.layers.58.self_attn.v_proj.weight": "model-00118-of-00162.safetensors",
503
+ "model.layers.59.input_layernorm.weight": "model-00121-of-00162.safetensors",
504
+ "model.layers.59.mlp.down_proj.weight": "model-00121-of-00162.safetensors",
505
+ "model.layers.59.mlp.gate_proj.weight": "model-00120-of-00162.safetensors",
506
+ "model.layers.59.mlp.up_proj.weight": "model-00121-of-00162.safetensors",
507
+ "model.layers.59.post_attention_layernorm.weight": "model-00121-of-00162.safetensors",
508
+ "model.layers.59.self_attn.k_proj.weight": "model-00120-of-00162.safetensors",
509
+ "model.layers.59.self_attn.o_proj.weight": "model-00120-of-00162.safetensors",
510
+ "model.layers.59.self_attn.q_proj.weight": "model-00120-of-00162.safetensors",
511
+ "model.layers.59.self_attn.v_proj.weight": "model-00120-of-00162.safetensors",
512
+ "model.layers.6.input_layernorm.weight": "model-00015-of-00162.safetensors",
513
+ "model.layers.6.mlp.down_proj.weight": "model-00015-of-00162.safetensors",
514
+ "model.layers.6.mlp.gate_proj.weight": "model-00014-of-00162.safetensors",
515
+ "model.layers.6.mlp.up_proj.weight": "model-00015-of-00162.safetensors",
516
+ "model.layers.6.post_attention_layernorm.weight": "model-00015-of-00162.safetensors",
517
+ "model.layers.6.self_attn.k_proj.weight": "model-00014-of-00162.safetensors",
518
+ "model.layers.6.self_attn.o_proj.weight": "model-00014-of-00162.safetensors",
519
+ "model.layers.6.self_attn.q_proj.weight": "model-00014-of-00162.safetensors",
520
+ "model.layers.6.self_attn.v_proj.weight": "model-00014-of-00162.safetensors",
521
+ "model.layers.60.input_layernorm.weight": "model-00123-of-00162.safetensors",
522
+ "model.layers.60.mlp.down_proj.weight": "model-00123-of-00162.safetensors",
523
+ "model.layers.60.mlp.gate_proj.weight": "model-00122-of-00162.safetensors",
524
+ "model.layers.60.mlp.up_proj.weight": "model-00123-of-00162.safetensors",
525
+ "model.layers.60.post_attention_layernorm.weight": "model-00123-of-00162.safetensors",
526
+ "model.layers.60.self_attn.k_proj.weight": "model-00122-of-00162.safetensors",
527
+ "model.layers.60.self_attn.o_proj.weight": "model-00122-of-00162.safetensors",
528
+ "model.layers.60.self_attn.q_proj.weight": "model-00122-of-00162.safetensors",
529
+ "model.layers.60.self_attn.v_proj.weight": "model-00122-of-00162.safetensors",
530
+ "model.layers.61.input_layernorm.weight": "model-00125-of-00162.safetensors",
531
+ "model.layers.61.mlp.down_proj.weight": "model-00125-of-00162.safetensors",
532
+ "model.layers.61.mlp.gate_proj.weight": "model-00124-of-00162.safetensors",
533
+ "model.layers.61.mlp.up_proj.weight": "model-00125-of-00162.safetensors",
534
+ "model.layers.61.post_attention_layernorm.weight": "model-00125-of-00162.safetensors",
535
+ "model.layers.61.self_attn.k_proj.weight": "model-00124-of-00162.safetensors",
536
+ "model.layers.61.self_attn.o_proj.weight": "model-00124-of-00162.safetensors",
537
+ "model.layers.61.self_attn.q_proj.weight": "model-00124-of-00162.safetensors",
538
+ "model.layers.61.self_attn.v_proj.weight": "model-00124-of-00162.safetensors",
539
+ "model.layers.62.input_layernorm.weight": "model-00127-of-00162.safetensors",
540
+ "model.layers.62.mlp.down_proj.weight": "model-00127-of-00162.safetensors",
541
+ "model.layers.62.mlp.gate_proj.weight": "model-00126-of-00162.safetensors",
542
+ "model.layers.62.mlp.up_proj.weight": "model-00127-of-00162.safetensors",
543
+ "model.layers.62.post_attention_layernorm.weight": "model-00127-of-00162.safetensors",
544
+ "model.layers.62.self_attn.k_proj.weight": "model-00126-of-00162.safetensors",
545
+ "model.layers.62.self_attn.o_proj.weight": "model-00126-of-00162.safetensors",
546
+ "model.layers.62.self_attn.q_proj.weight": "model-00126-of-00162.safetensors",
547
+ "model.layers.62.self_attn.v_proj.weight": "model-00126-of-00162.safetensors",
548
+ "model.layers.63.input_layernorm.weight": "model-00129-of-00162.safetensors",
549
+ "model.layers.63.mlp.down_proj.weight": "model-00129-of-00162.safetensors",
550
+ "model.layers.63.mlp.gate_proj.weight": "model-00128-of-00162.safetensors",
551
+ "model.layers.63.mlp.up_proj.weight": "model-00129-of-00162.safetensors",
552
+ "model.layers.63.post_attention_layernorm.weight": "model-00129-of-00162.safetensors",
553
+ "model.layers.63.self_attn.k_proj.weight": "model-00128-of-00162.safetensors",
554
+ "model.layers.63.self_attn.o_proj.weight": "model-00128-of-00162.safetensors",
555
+ "model.layers.63.self_attn.q_proj.weight": "model-00128-of-00162.safetensors",
556
+ "model.layers.63.self_attn.v_proj.weight": "model-00128-of-00162.safetensors",
557
+ "model.layers.64.input_layernorm.weight": "model-00131-of-00162.safetensors",
558
+ "model.layers.64.mlp.down_proj.weight": "model-00131-of-00162.safetensors",
559
+ "model.layers.64.mlp.gate_proj.weight": "model-00130-of-00162.safetensors",
560
+ "model.layers.64.mlp.up_proj.weight": "model-00131-of-00162.safetensors",
561
+ "model.layers.64.post_attention_layernorm.weight": "model-00131-of-00162.safetensors",
562
+ "model.layers.64.self_attn.k_proj.weight": "model-00130-of-00162.safetensors",
563
+ "model.layers.64.self_attn.o_proj.weight": "model-00130-of-00162.safetensors",
564
+ "model.layers.64.self_attn.q_proj.weight": "model-00130-of-00162.safetensors",
565
+ "model.layers.64.self_attn.v_proj.weight": "model-00130-of-00162.safetensors",
566
+ "model.layers.65.input_layernorm.weight": "model-00133-of-00162.safetensors",
567
+ "model.layers.65.mlp.down_proj.weight": "model-00133-of-00162.safetensors",
568
+ "model.layers.65.mlp.gate_proj.weight": "model-00132-of-00162.safetensors",
569
+ "model.layers.65.mlp.up_proj.weight": "model-00133-of-00162.safetensors",
570
+ "model.layers.65.post_attention_layernorm.weight": "model-00133-of-00162.safetensors",
571
+ "model.layers.65.self_attn.k_proj.weight": "model-00132-of-00162.safetensors",
572
+ "model.layers.65.self_attn.o_proj.weight": "model-00132-of-00162.safetensors",
573
+ "model.layers.65.self_attn.q_proj.weight": "model-00132-of-00162.safetensors",
574
+ "model.layers.65.self_attn.v_proj.weight": "model-00132-of-00162.safetensors",
575
+ "model.layers.66.input_layernorm.weight": "model-00135-of-00162.safetensors",
576
+ "model.layers.66.mlp.down_proj.weight": "model-00135-of-00162.safetensors",
577
+ "model.layers.66.mlp.gate_proj.weight": "model-00134-of-00162.safetensors",
578
+ "model.layers.66.mlp.up_proj.weight": "model-00135-of-00162.safetensors",
579
+ "model.layers.66.post_attention_layernorm.weight": "model-00135-of-00162.safetensors",
580
+ "model.layers.66.self_attn.k_proj.weight": "model-00134-of-00162.safetensors",
581
+ "model.layers.66.self_attn.o_proj.weight": "model-00134-of-00162.safetensors",
582
+ "model.layers.66.self_attn.q_proj.weight": "model-00134-of-00162.safetensors",
583
+ "model.layers.66.self_attn.v_proj.weight": "model-00134-of-00162.safetensors",
584
+ "model.layers.67.input_layernorm.weight": "model-00137-of-00162.safetensors",
585
+ "model.layers.67.mlp.down_proj.weight": "model-00137-of-00162.safetensors",
586
+ "model.layers.67.mlp.gate_proj.weight": "model-00136-of-00162.safetensors",
587
+ "model.layers.67.mlp.up_proj.weight": "model-00137-of-00162.safetensors",
588
+ "model.layers.67.post_attention_layernorm.weight": "model-00137-of-00162.safetensors",
589
+ "model.layers.67.self_attn.k_proj.weight": "model-00136-of-00162.safetensors",
590
+ "model.layers.67.self_attn.o_proj.weight": "model-00136-of-00162.safetensors",
591
+ "model.layers.67.self_attn.q_proj.weight": "model-00136-of-00162.safetensors",
592
+ "model.layers.67.self_attn.v_proj.weight": "model-00136-of-00162.safetensors",
593
+ "model.layers.68.input_layernorm.weight": "model-00139-of-00162.safetensors",
594
+ "model.layers.68.mlp.down_proj.weight": "model-00139-of-00162.safetensors",
595
+ "model.layers.68.mlp.gate_proj.weight": "model-00138-of-00162.safetensors",
596
+ "model.layers.68.mlp.up_proj.weight": "model-00139-of-00162.safetensors",
597
+ "model.layers.68.post_attention_layernorm.weight": "model-00139-of-00162.safetensors",
598
+ "model.layers.68.self_attn.k_proj.weight": "model-00138-of-00162.safetensors",
599
+ "model.layers.68.self_attn.o_proj.weight": "model-00138-of-00162.safetensors",
600
+ "model.layers.68.self_attn.q_proj.weight": "model-00138-of-00162.safetensors",
601
+ "model.layers.68.self_attn.v_proj.weight": "model-00138-of-00162.safetensors",
602
+ "model.layers.69.input_layernorm.weight": "model-00141-of-00162.safetensors",
603
+ "model.layers.69.mlp.down_proj.weight": "model-00141-of-00162.safetensors",
604
+ "model.layers.69.mlp.gate_proj.weight": "model-00140-of-00162.safetensors",
605
+ "model.layers.69.mlp.up_proj.weight": "model-00141-of-00162.safetensors",
606
+ "model.layers.69.post_attention_layernorm.weight": "model-00141-of-00162.safetensors",
607
+ "model.layers.69.self_attn.k_proj.weight": "model-00140-of-00162.safetensors",
608
+ "model.layers.69.self_attn.o_proj.weight": "model-00140-of-00162.safetensors",
609
+ "model.layers.69.self_attn.q_proj.weight": "model-00140-of-00162.safetensors",
610
+ "model.layers.69.self_attn.v_proj.weight": "model-00140-of-00162.safetensors",
611
+ "model.layers.7.input_layernorm.weight": "model-00017-of-00162.safetensors",
612
+ "model.layers.7.mlp.down_proj.weight": "model-00017-of-00162.safetensors",
613
+ "model.layers.7.mlp.gate_proj.weight": "model-00016-of-00162.safetensors",
614
+ "model.layers.7.mlp.up_proj.weight": "model-00017-of-00162.safetensors",
615
+ "model.layers.7.post_attention_layernorm.weight": "model-00017-of-00162.safetensors",
616
+ "model.layers.7.self_attn.k_proj.weight": "model-00016-of-00162.safetensors",
617
+ "model.layers.7.self_attn.o_proj.weight": "model-00016-of-00162.safetensors",
618
+ "model.layers.7.self_attn.q_proj.weight": "model-00016-of-00162.safetensors",
619
+ "model.layers.7.self_attn.v_proj.weight": "model-00016-of-00162.safetensors",
620
+ "model.layers.70.input_layernorm.weight": "model-00143-of-00162.safetensors",
621
+ "model.layers.70.mlp.down_proj.weight": "model-00143-of-00162.safetensors",
622
+ "model.layers.70.mlp.gate_proj.weight": "model-00142-of-00162.safetensors",
623
+ "model.layers.70.mlp.up_proj.weight": "model-00143-of-00162.safetensors",
624
+ "model.layers.70.post_attention_layernorm.weight": "model-00143-of-00162.safetensors",
625
+ "model.layers.70.self_attn.k_proj.weight": "model-00142-of-00162.safetensors",
626
+ "model.layers.70.self_attn.o_proj.weight": "model-00142-of-00162.safetensors",
627
+ "model.layers.70.self_attn.q_proj.weight": "model-00142-of-00162.safetensors",
628
+ "model.layers.70.self_attn.v_proj.weight": "model-00142-of-00162.safetensors",
629
+ "model.layers.71.input_layernorm.weight": "model-00145-of-00162.safetensors",
630
+ "model.layers.71.mlp.down_proj.weight": "model-00145-of-00162.safetensors",
631
+ "model.layers.71.mlp.gate_proj.weight": "model-00144-of-00162.safetensors",
632
+ "model.layers.71.mlp.up_proj.weight": "model-00145-of-00162.safetensors",
633
+ "model.layers.71.post_attention_layernorm.weight": "model-00145-of-00162.safetensors",
634
+ "model.layers.71.self_attn.k_proj.weight": "model-00144-of-00162.safetensors",
635
+ "model.layers.71.self_attn.o_proj.weight": "model-00144-of-00162.safetensors",
636
+ "model.layers.71.self_attn.q_proj.weight": "model-00144-of-00162.safetensors",
637
+ "model.layers.71.self_attn.v_proj.weight": "model-00144-of-00162.safetensors",
638
+ "model.layers.72.input_layernorm.weight": "model-00147-of-00162.safetensors",
639
+ "model.layers.72.mlp.down_proj.weight": "model-00147-of-00162.safetensors",
640
+ "model.layers.72.mlp.gate_proj.weight": "model-00146-of-00162.safetensors",
641
+ "model.layers.72.mlp.up_proj.weight": "model-00147-of-00162.safetensors",
642
+ "model.layers.72.post_attention_layernorm.weight": "model-00147-of-00162.safetensors",
643
+ "model.layers.72.self_attn.k_proj.weight": "model-00146-of-00162.safetensors",
644
+ "model.layers.72.self_attn.o_proj.weight": "model-00146-of-00162.safetensors",
645
+ "model.layers.72.self_attn.q_proj.weight": "model-00146-of-00162.safetensors",
646
+ "model.layers.72.self_attn.v_proj.weight": "model-00146-of-00162.safetensors",
647
+ "model.layers.73.input_layernorm.weight": "model-00149-of-00162.safetensors",
648
+ "model.layers.73.mlp.down_proj.weight": "model-00149-of-00162.safetensors",
649
+ "model.layers.73.mlp.gate_proj.weight": "model-00148-of-00162.safetensors",
650
+ "model.layers.73.mlp.up_proj.weight": "model-00149-of-00162.safetensors",
651
+ "model.layers.73.post_attention_layernorm.weight": "model-00149-of-00162.safetensors",
652
+ "model.layers.73.self_attn.k_proj.weight": "model-00148-of-00162.safetensors",
653
+ "model.layers.73.self_attn.o_proj.weight": "model-00148-of-00162.safetensors",
654
+ "model.layers.73.self_attn.q_proj.weight": "model-00148-of-00162.safetensors",
655
+ "model.layers.73.self_attn.v_proj.weight": "model-00148-of-00162.safetensors",
656
+ "model.layers.74.input_layernorm.weight": "model-00151-of-00162.safetensors",
657
+ "model.layers.74.mlp.down_proj.weight": "model-00151-of-00162.safetensors",
658
+ "model.layers.74.mlp.gate_proj.weight": "model-00150-of-00162.safetensors",
659
+ "model.layers.74.mlp.up_proj.weight": "model-00151-of-00162.safetensors",
660
+ "model.layers.74.post_attention_layernorm.weight": "model-00151-of-00162.safetensors",
661
+ "model.layers.74.self_attn.k_proj.weight": "model-00150-of-00162.safetensors",
662
+ "model.layers.74.self_attn.o_proj.weight": "model-00150-of-00162.safetensors",
663
+ "model.layers.74.self_attn.q_proj.weight": "model-00150-of-00162.safetensors",
664
+ "model.layers.74.self_attn.v_proj.weight": "model-00150-of-00162.safetensors",
665
+ "model.layers.75.input_layernorm.weight": "model-00153-of-00162.safetensors",
666
+ "model.layers.75.mlp.down_proj.weight": "model-00153-of-00162.safetensors",
667
+ "model.layers.75.mlp.gate_proj.weight": "model-00152-of-00162.safetensors",
668
+ "model.layers.75.mlp.up_proj.weight": "model-00153-of-00162.safetensors",
669
+ "model.layers.75.post_attention_layernorm.weight": "model-00153-of-00162.safetensors",
670
+ "model.layers.75.self_attn.k_proj.weight": "model-00152-of-00162.safetensors",
671
+ "model.layers.75.self_attn.o_proj.weight": "model-00152-of-00162.safetensors",
672
+ "model.layers.75.self_attn.q_proj.weight": "model-00152-of-00162.safetensors",
673
+ "model.layers.75.self_attn.v_proj.weight": "model-00152-of-00162.safetensors",
674
+ "model.layers.76.input_layernorm.weight": "model-00155-of-00162.safetensors",
675
+ "model.layers.76.mlp.down_proj.weight": "model-00155-of-00162.safetensors",
676
+ "model.layers.76.mlp.gate_proj.weight": "model-00154-of-00162.safetensors",
677
+ "model.layers.76.mlp.up_proj.weight": "model-00155-of-00162.safetensors",
678
+ "model.layers.76.post_attention_layernorm.weight": "model-00155-of-00162.safetensors",
679
+ "model.layers.76.self_attn.k_proj.weight": "model-00154-of-00162.safetensors",
680
+ "model.layers.76.self_attn.o_proj.weight": "model-00154-of-00162.safetensors",
681
+ "model.layers.76.self_attn.q_proj.weight": "model-00154-of-00162.safetensors",
682
+ "model.layers.76.self_attn.v_proj.weight": "model-00154-of-00162.safetensors",
683
+ "model.layers.77.input_layernorm.weight": "model-00157-of-00162.safetensors",
684
+ "model.layers.77.mlp.down_proj.weight": "model-00157-of-00162.safetensors",
685
+ "model.layers.77.mlp.gate_proj.weight": "model-00156-of-00162.safetensors",
686
+ "model.layers.77.mlp.up_proj.weight": "model-00157-of-00162.safetensors",
687
+ "model.layers.77.post_attention_layernorm.weight": "model-00157-of-00162.safetensors",
688
+ "model.layers.77.self_attn.k_proj.weight": "model-00156-of-00162.safetensors",
689
+ "model.layers.77.self_attn.o_proj.weight": "model-00156-of-00162.safetensors",
690
+ "model.layers.77.self_attn.q_proj.weight": "model-00156-of-00162.safetensors",
691
+ "model.layers.77.self_attn.v_proj.weight": "model-00156-of-00162.safetensors",
692
+ "model.layers.78.input_layernorm.weight": "model-00159-of-00162.safetensors",
693
+ "model.layers.78.mlp.down_proj.weight": "model-00159-of-00162.safetensors",
694
+ "model.layers.78.mlp.gate_proj.weight": "model-00158-of-00162.safetensors",
695
+ "model.layers.78.mlp.up_proj.weight": "model-00159-of-00162.safetensors",
696
+ "model.layers.78.post_attention_layernorm.weight": "model-00159-of-00162.safetensors",
697
+ "model.layers.78.self_attn.k_proj.weight": "model-00158-of-00162.safetensors",
698
+ "model.layers.78.self_attn.o_proj.weight": "model-00158-of-00162.safetensors",
699
+ "model.layers.78.self_attn.q_proj.weight": "model-00158-of-00162.safetensors",
700
+ "model.layers.78.self_attn.v_proj.weight": "model-00158-of-00162.safetensors",
701
+ "model.layers.79.input_layernorm.weight": "model-00162-of-00162.safetensors",
702
+ "model.layers.79.mlp.down_proj.weight": "model-00162-of-00162.safetensors",
703
+ "model.layers.79.mlp.gate_proj.weight": "model-00160-of-00162.safetensors",
704
+ "model.layers.79.mlp.up_proj.weight": "model-00162-of-00162.safetensors",
705
+ "model.layers.79.post_attention_layernorm.weight": "model-00162-of-00162.safetensors",
706
+ "model.layers.79.self_attn.k_proj.weight": "model-00160-of-00162.safetensors",
707
+ "model.layers.79.self_attn.o_proj.weight": "model-00160-of-00162.safetensors",
708
+ "model.layers.79.self_attn.q_proj.weight": "model-00160-of-00162.safetensors",
709
+ "model.layers.79.self_attn.v_proj.weight": "model-00160-of-00162.safetensors",
710
+ "model.layers.8.input_layernorm.weight": "model-00019-of-00162.safetensors",
711
+ "model.layers.8.mlp.down_proj.weight": "model-00019-of-00162.safetensors",
712
+ "model.layers.8.mlp.gate_proj.weight": "model-00018-of-00162.safetensors",
713
+ "model.layers.8.mlp.up_proj.weight": "model-00019-of-00162.safetensors",
714
+ "model.layers.8.post_attention_layernorm.weight": "model-00019-of-00162.safetensors",
715
+ "model.layers.8.self_attn.k_proj.weight": "model-00018-of-00162.safetensors",
716
+ "model.layers.8.self_attn.o_proj.weight": "model-00018-of-00162.safetensors",
717
+ "model.layers.8.self_attn.q_proj.weight": "model-00018-of-00162.safetensors",
718
+ "model.layers.8.self_attn.v_proj.weight": "model-00018-of-00162.safetensors",
719
+ "model.layers.9.input_layernorm.weight": "model-00021-of-00162.safetensors",
720
+ "model.layers.9.mlp.down_proj.weight": "model-00021-of-00162.safetensors",
721
+ "model.layers.9.mlp.gate_proj.weight": "model-00020-of-00162.safetensors",
722
+ "model.layers.9.mlp.up_proj.weight": "model-00021-of-00162.safetensors",
723
+ "model.layers.9.post_attention_layernorm.weight": "model-00021-of-00162.safetensors",
724
+ "model.layers.9.self_attn.k_proj.weight": "model-00020-of-00162.safetensors",
725
+ "model.layers.9.self_attn.o_proj.weight": "model-00020-of-00162.safetensors",
726
+ "model.layers.9.self_attn.q_proj.weight": "model-00020-of-00162.safetensors",
727
+ "model.layers.9.self_attn.v_proj.weight": "model-00020-of-00162.safetensors",
728
+ "model.norm.weight": "model-00162-of-00162.safetensors"
729
+ }
730
+ }
output-00001-of-00008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da353760a5e0b0420ad14b2281ce144bfd36a93e14a6c1564f43d7a1c365b974
3
+ size 8476752432
output-00002-of-00008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9959b53544eb25e2d1f1fbf2d08a8b0497ca6d7db4a832834e9d24adbefd3792
3
+ size 8542205356
output-00003-of-00008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6065ed3b0281044d100af9e362dfde3896c1ea1c8197cbe9140fc8fd25041e59
3
+ size 8500730756
output-00004-of-00008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3b6362ffde3bc14d70652bf905c14fa161650800fa8d8e15385929cfe98171a
3
+ size 8501450752
output-00005-of-00008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:614a597b3bcf01328e93299c0971b4f239bcbc9b884f467b91638ecb1c70f293
3
+ size 8417999540
output-00006-of-00008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3fd596207a1d9dac2694bbc75a17af0f06853bb2f10283c96f02601af263602
3
+ size 8371780624
output-00007-of-00008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53d384ea92676acef127bbdb093baafc6098ef2d77889655790304ed39743336
3
+ size 8465390608
output-00008-of-00008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2de2c0b7e39d730d799b88007b357e6465c426ae2286c08cc0a5898bc0f15ee6
3
+ size 6411824792
special_tokens_map.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<thinking>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "</thinking>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<output>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "</output>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ {
32
+ "content": "<reflection>",
33
+ "lstrip": false,
34
+ "normalized": false,
35
+ "rstrip": false,
36
+ "single_word": false
37
+ },
38
+ {
39
+ "content": "</reflection>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ }
45
+ ],
46
+ "bos_token": {
47
+ "content": "<|begin_of_text|>",
48
+ "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": false,
51
+ "single_word": false
52
+ },
53
+ "eos_token": {
54
+ "content": "<|eot_id|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false
59
+ },
60
+ "pad_token": "<|eot_id|>"
61
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,2119 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "128000": {
4
+ "content": "<|begin_of_text|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "128001": {
12
+ "content": "<|end_of_text|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "128002": {
20
+ "content": "<|reserved_special_token_0|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "128003": {
28
+ "content": "<|reserved_special_token_1|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128004": {
36
+ "content": "<|finetune_right_pad_id|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "128005": {
44
+ "content": "<|reserved_special_token_2|>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "128006": {
52
+ "content": "<|start_header_id|>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "128007": {
60
+ "content": "<|end_header_id|>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "128008": {
68
+ "content": "<|eom_id|>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "128009": {
76
+ "content": "<|eot_id|>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "128010": {
84
+ "content": "<|python_tag|>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "128011": {
92
+ "content": "<|reserved_special_token_3|>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "128012": {
100
+ "content": "<|reserved_special_token_4|>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "128013": {
108
+ "content": "<|reserved_special_token_5|>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "128014": {
116
+ "content": "<|reserved_special_token_6|>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "128015": {
124
+ "content": "<|reserved_special_token_7|>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "128016": {
132
+ "content": "<|reserved_special_token_8|>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "128017": {
140
+ "content": "<|reserved_special_token_9|>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "128018": {
148
+ "content": "<|reserved_special_token_10|>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "128019": {
156
+ "content": "<|reserved_special_token_11|>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "128020": {
164
+ "content": "<|reserved_special_token_12|>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "128021": {
172
+ "content": "<|reserved_special_token_13|>",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": true
178
+ },
179
+ "128022": {
180
+ "content": "<|reserved_special_token_14|>",
181
+ "lstrip": false,
182
+ "normalized": false,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": true
186
+ },
187
+ "128023": {
188
+ "content": "<|reserved_special_token_15|>",
189
+ "lstrip": false,
190
+ "normalized": false,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": true
194
+ },
195
+ "128024": {
196
+ "content": "<|reserved_special_token_16|>",
197
+ "lstrip": false,
198
+ "normalized": false,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": true
202
+ },
203
+ "128025": {
204
+ "content": "<|reserved_special_token_17|>",
205
+ "lstrip": false,
206
+ "normalized": false,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": true
210
+ },
211
+ "128026": {
212
+ "content": "<|reserved_special_token_18|>",
213
+ "lstrip": false,
214
+ "normalized": false,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": true
218
+ },
219
+ "128027": {
220
+ "content": "<|reserved_special_token_19|>",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "128028": {
228
+ "content": "<|reserved_special_token_20|>",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "128029": {
236
+ "content": "<|reserved_special_token_21|>",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "128030": {
244
+ "content": "<|reserved_special_token_22|>",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "128031": {
252
+ "content": "<|reserved_special_token_23|>",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "128032": {
260
+ "content": "<|reserved_special_token_24|>",
261
+ "lstrip": false,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "128033": {
268
+ "content": "<|reserved_special_token_25|>",
269
+ "lstrip": false,
270
+ "normalized": false,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": true
274
+ },
275
+ "128034": {
276
+ "content": "<|reserved_special_token_26|>",
277
+ "lstrip": false,
278
+ "normalized": false,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": true
282
+ },
283
+ "128035": {
284
+ "content": "<|reserved_special_token_27|>",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": true
290
+ },
291
+ "128036": {
292
+ "content": "<|reserved_special_token_28|>",
293
+ "lstrip": false,
294
+ "normalized": false,
295
+ "rstrip": false,
296
+ "single_word": false,
297
+ "special": true
298
+ },
299
+ "128037": {
300
+ "content": "<|reserved_special_token_29|>",
301
+ "lstrip": false,
302
+ "normalized": false,
303
+ "rstrip": false,
304
+ "single_word": false,
305
+ "special": true
306
+ },
307
+ "128038": {
308
+ "content": "<|reserved_special_token_30|>",
309
+ "lstrip": false,
310
+ "normalized": false,
311
+ "rstrip": false,
312
+ "single_word": false,
313
+ "special": true
314
+ },
315
+ "128039": {
316
+ "content": "<|reserved_special_token_31|>",
317
+ "lstrip": false,
318
+ "normalized": false,
319
+ "rstrip": false,
320
+ "single_word": false,
321
+ "special": true
322
+ },
323
+ "128040": {
324
+ "content": "<|reserved_special_token_32|>",
325
+ "lstrip": false,
326
+ "normalized": false,
327
+ "rstrip": false,
328
+ "single_word": false,
329
+ "special": true
330
+ },
331
+ "128041": {
332
+ "content": "<|reserved_special_token_33|>",
333
+ "lstrip": false,
334
+ "normalized": false,
335
+ "rstrip": false,
336
+ "single_word": false,
337
+ "special": true
338
+ },
339
+ "128042": {
340
+ "content": "<|reserved_special_token_34|>",
341
+ "lstrip": false,
342
+ "normalized": false,
343
+ "rstrip": false,
344
+ "single_word": false,
345
+ "special": true
346
+ },
347
+ "128043": {
348
+ "content": "<|reserved_special_token_35|>",
349
+ "lstrip": false,
350
+ "normalized": false,
351
+ "rstrip": false,
352
+ "single_word": false,
353
+ "special": true
354
+ },
355
+ "128044": {
356
+ "content": "<|reserved_special_token_36|>",
357
+ "lstrip": false,
358
+ "normalized": false,
359
+ "rstrip": false,
360
+ "single_word": false,
361
+ "special": true
362
+ },
363
+ "128045": {
364
+ "content": "<|reserved_special_token_37|>",
365
+ "lstrip": false,
366
+ "normalized": false,
367
+ "rstrip": false,
368
+ "single_word": false,
369
+ "special": true
370
+ },
371
+ "128046": {
372
+ "content": "<|reserved_special_token_38|>",
373
+ "lstrip": false,
374
+ "normalized": false,
375
+ "rstrip": false,
376
+ "single_word": false,
377
+ "special": true
378
+ },
379
+ "128047": {
380
+ "content": "<|reserved_special_token_39|>",
381
+ "lstrip": false,
382
+ "normalized": false,
383
+ "rstrip": false,
384
+ "single_word": false,
385
+ "special": true
386
+ },
387
+ "128048": {
388
+ "content": "<|reserved_special_token_40|>",
389
+ "lstrip": false,
390
+ "normalized": false,
391
+ "rstrip": false,
392
+ "single_word": false,
393
+ "special": true
394
+ },
395
+ "128049": {
396
+ "content": "<|reserved_special_token_41|>",
397
+ "lstrip": false,
398
+ "normalized": false,
399
+ "rstrip": false,
400
+ "single_word": false,
401
+ "special": true
402
+ },
403
+ "128050": {
404
+ "content": "<|reserved_special_token_42|>",
405
+ "lstrip": false,
406
+ "normalized": false,
407
+ "rstrip": false,
408
+ "single_word": false,
409
+ "special": true
410
+ },
411
+ "128051": {
412
+ "content": "<|reserved_special_token_43|>",
413
+ "lstrip": false,
414
+ "normalized": false,
415
+ "rstrip": false,
416
+ "single_word": false,
417
+ "special": true
418
+ },
419
+ "128052": {
420
+ "content": "<|reserved_special_token_44|>",
421
+ "lstrip": false,
422
+ "normalized": false,
423
+ "rstrip": false,
424
+ "single_word": false,
425
+ "special": true
426
+ },
427
+ "128053": {
428
+ "content": "<|reserved_special_token_45|>",
429
+ "lstrip": false,
430
+ "normalized": false,
431
+ "rstrip": false,
432
+ "single_word": false,
433
+ "special": true
434
+ },
435
+ "128054": {
436
+ "content": "<|reserved_special_token_46|>",
437
+ "lstrip": false,
438
+ "normalized": false,
439
+ "rstrip": false,
440
+ "single_word": false,
441
+ "special": true
442
+ },
443
+ "128055": {
444
+ "content": "<|reserved_special_token_47|>",
445
+ "lstrip": false,
446
+ "normalized": false,
447
+ "rstrip": false,
448
+ "single_word": false,
449
+ "special": true
450
+ },
451
+ "128056": {
452
+ "content": "<|reserved_special_token_48|>",
453
+ "lstrip": false,
454
+ "normalized": false,
455
+ "rstrip": false,
456
+ "single_word": false,
457
+ "special": true
458
+ },
459
+ "128057": {
460
+ "content": "<|reserved_special_token_49|>",
461
+ "lstrip": false,
462
+ "normalized": false,
463
+ "rstrip": false,
464
+ "single_word": false,
465
+ "special": true
466
+ },
467
+ "128058": {
468
+ "content": "<|reserved_special_token_50|>",
469
+ "lstrip": false,
470
+ "normalized": false,
471
+ "rstrip": false,
472
+ "single_word": false,
473
+ "special": true
474
+ },
475
+ "128059": {
476
+ "content": "<|reserved_special_token_51|>",
477
+ "lstrip": false,
478
+ "normalized": false,
479
+ "rstrip": false,
480
+ "single_word": false,
481
+ "special": true
482
+ },
483
+ "128060": {
484
+ "content": "<|reserved_special_token_52|>",
485
+ "lstrip": false,
486
+ "normalized": false,
487
+ "rstrip": false,
488
+ "single_word": false,
489
+ "special": true
490
+ },
491
+ "128061": {
492
+ "content": "<|reserved_special_token_53|>",
493
+ "lstrip": false,
494
+ "normalized": false,
495
+ "rstrip": false,
496
+ "single_word": false,
497
+ "special": true
498
+ },
499
+ "128062": {
500
+ "content": "<|reserved_special_token_54|>",
501
+ "lstrip": false,
502
+ "normalized": false,
503
+ "rstrip": false,
504
+ "single_word": false,
505
+ "special": true
506
+ },
507
+ "128063": {
508
+ "content": "<|reserved_special_token_55|>",
509
+ "lstrip": false,
510
+ "normalized": false,
511
+ "rstrip": false,
512
+ "single_word": false,
513
+ "special": true
514
+ },
515
+ "128064": {
516
+ "content": "<|reserved_special_token_56|>",
517
+ "lstrip": false,
518
+ "normalized": false,
519
+ "rstrip": false,
520
+ "single_word": false,
521
+ "special": true
522
+ },
523
+ "128065": {
524
+ "content": "<|reserved_special_token_57|>",
525
+ "lstrip": false,
526
+ "normalized": false,
527
+ "rstrip": false,
528
+ "single_word": false,
529
+ "special": true
530
+ },
531
+ "128066": {
532
+ "content": "<|reserved_special_token_58|>",
533
+ "lstrip": false,
534
+ "normalized": false,
535
+ "rstrip": false,
536
+ "single_word": false,
537
+ "special": true
538
+ },
539
+ "128067": {
540
+ "content": "<|reserved_special_token_59|>",
541
+ "lstrip": false,
542
+ "normalized": false,
543
+ "rstrip": false,
544
+ "single_word": false,
545
+ "special": true
546
+ },
547
+ "128068": {
548
+ "content": "<|reserved_special_token_60|>",
549
+ "lstrip": false,
550
+ "normalized": false,
551
+ "rstrip": false,
552
+ "single_word": false,
553
+ "special": true
554
+ },
555
+ "128069": {
556
+ "content": "<|reserved_special_token_61|>",
557
+ "lstrip": false,
558
+ "normalized": false,
559
+ "rstrip": false,
560
+ "single_word": false,
561
+ "special": true
562
+ },
563
+ "128070": {
564
+ "content": "<|reserved_special_token_62|>",
565
+ "lstrip": false,
566
+ "normalized": false,
567
+ "rstrip": false,
568
+ "single_word": false,
569
+ "special": true
570
+ },
571
+ "128071": {
572
+ "content": "<|reserved_special_token_63|>",
573
+ "lstrip": false,
574
+ "normalized": false,
575
+ "rstrip": false,
576
+ "single_word": false,
577
+ "special": true
578
+ },
579
+ "128072": {
580
+ "content": "<|reserved_special_token_64|>",
581
+ "lstrip": false,
582
+ "normalized": false,
583
+ "rstrip": false,
584
+ "single_word": false,
585
+ "special": true
586
+ },
587
+ "128073": {
588
+ "content": "<|reserved_special_token_65|>",
589
+ "lstrip": false,
590
+ "normalized": false,
591
+ "rstrip": false,
592
+ "single_word": false,
593
+ "special": true
594
+ },
595
+ "128074": {
596
+ "content": "<|reserved_special_token_66|>",
597
+ "lstrip": false,
598
+ "normalized": false,
599
+ "rstrip": false,
600
+ "single_word": false,
601
+ "special": true
602
+ },
603
+ "128075": {
604
+ "content": "<|reserved_special_token_67|>",
605
+ "lstrip": false,
606
+ "normalized": false,
607
+ "rstrip": false,
608
+ "single_word": false,
609
+ "special": true
610
+ },
611
+ "128076": {
612
+ "content": "<|reserved_special_token_68|>",
613
+ "lstrip": false,
614
+ "normalized": false,
615
+ "rstrip": false,
616
+ "single_word": false,
617
+ "special": true
618
+ },
619
+ "128077": {
620
+ "content": "<|reserved_special_token_69|>",
621
+ "lstrip": false,
622
+ "normalized": false,
623
+ "rstrip": false,
624
+ "single_word": false,
625
+ "special": true
626
+ },
627
+ "128078": {
628
+ "content": "<|reserved_special_token_70|>",
629
+ "lstrip": false,
630
+ "normalized": false,
631
+ "rstrip": false,
632
+ "single_word": false,
633
+ "special": true
634
+ },
635
+ "128079": {
636
+ "content": "<|reserved_special_token_71|>",
637
+ "lstrip": false,
638
+ "normalized": false,
639
+ "rstrip": false,
640
+ "single_word": false,
641
+ "special": true
642
+ },
643
+ "128080": {
644
+ "content": "<|reserved_special_token_72|>",
645
+ "lstrip": false,
646
+ "normalized": false,
647
+ "rstrip": false,
648
+ "single_word": false,
649
+ "special": true
650
+ },
651
+ "128081": {
652
+ "content": "<|reserved_special_token_73|>",
653
+ "lstrip": false,
654
+ "normalized": false,
655
+ "rstrip": false,
656
+ "single_word": false,
657
+ "special": true
658
+ },
659
+ "128082": {
660
+ "content": "<|reserved_special_token_74|>",
661
+ "lstrip": false,
662
+ "normalized": false,
663
+ "rstrip": false,
664
+ "single_word": false,
665
+ "special": true
666
+ },
667
+ "128083": {
668
+ "content": "<|reserved_special_token_75|>",
669
+ "lstrip": false,
670
+ "normalized": false,
671
+ "rstrip": false,
672
+ "single_word": false,
673
+ "special": true
674
+ },
675
+ "128084": {
676
+ "content": "<|reserved_special_token_76|>",
677
+ "lstrip": false,
678
+ "normalized": false,
679
+ "rstrip": false,
680
+ "single_word": false,
681
+ "special": true
682
+ },
683
+ "128085": {
684
+ "content": "<|reserved_special_token_77|>",
685
+ "lstrip": false,
686
+ "normalized": false,
687
+ "rstrip": false,
688
+ "single_word": false,
689
+ "special": true
690
+ },
691
+ "128086": {
692
+ "content": "<|reserved_special_token_78|>",
693
+ "lstrip": false,
694
+ "normalized": false,
695
+ "rstrip": false,
696
+ "single_word": false,
697
+ "special": true
698
+ },
699
+ "128087": {
700
+ "content": "<|reserved_special_token_79|>",
701
+ "lstrip": false,
702
+ "normalized": false,
703
+ "rstrip": false,
704
+ "single_word": false,
705
+ "special": true
706
+ },
707
+ "128088": {
708
+ "content": "<|reserved_special_token_80|>",
709
+ "lstrip": false,
710
+ "normalized": false,
711
+ "rstrip": false,
712
+ "single_word": false,
713
+ "special": true
714
+ },
715
+ "128089": {
716
+ "content": "<|reserved_special_token_81|>",
717
+ "lstrip": false,
718
+ "normalized": false,
719
+ "rstrip": false,
720
+ "single_word": false,
721
+ "special": true
722
+ },
723
+ "128090": {
724
+ "content": "<|reserved_special_token_82|>",
725
+ "lstrip": false,
726
+ "normalized": false,
727
+ "rstrip": false,
728
+ "single_word": false,
729
+ "special": true
730
+ },
731
+ "128091": {
732
+ "content": "<|reserved_special_token_83|>",
733
+ "lstrip": false,
734
+ "normalized": false,
735
+ "rstrip": false,
736
+ "single_word": false,
737
+ "special": true
738
+ },
739
+ "128092": {
740
+ "content": "<|reserved_special_token_84|>",
741
+ "lstrip": false,
742
+ "normalized": false,
743
+ "rstrip": false,
744
+ "single_word": false,
745
+ "special": true
746
+ },
747
+ "128093": {
748
+ "content": "<|reserved_special_token_85|>",
749
+ "lstrip": false,
750
+ "normalized": false,
751
+ "rstrip": false,
752
+ "single_word": false,
753
+ "special": true
754
+ },
755
+ "128094": {
756
+ "content": "<|reserved_special_token_86|>",
757
+ "lstrip": false,
758
+ "normalized": false,
759
+ "rstrip": false,
760
+ "single_word": false,
761
+ "special": true
762
+ },
763
+ "128095": {
764
+ "content": "<|reserved_special_token_87|>",
765
+ "lstrip": false,
766
+ "normalized": false,
767
+ "rstrip": false,
768
+ "single_word": false,
769
+ "special": true
770
+ },
771
+ "128096": {
772
+ "content": "<|reserved_special_token_88|>",
773
+ "lstrip": false,
774
+ "normalized": false,
775
+ "rstrip": false,
776
+ "single_word": false,
777
+ "special": true
778
+ },
779
+ "128097": {
780
+ "content": "<|reserved_special_token_89|>",
781
+ "lstrip": false,
782
+ "normalized": false,
783
+ "rstrip": false,
784
+ "single_word": false,
785
+ "special": true
786
+ },
787
+ "128098": {
788
+ "content": "<|reserved_special_token_90|>",
789
+ "lstrip": false,
790
+ "normalized": false,
791
+ "rstrip": false,
792
+ "single_word": false,
793
+ "special": true
794
+ },
795
+ "128099": {
796
+ "content": "<|reserved_special_token_91|>",
797
+ "lstrip": false,
798
+ "normalized": false,
799
+ "rstrip": false,
800
+ "single_word": false,
801
+ "special": true
802
+ },
803
+ "128100": {
804
+ "content": "<|reserved_special_token_92|>",
805
+ "lstrip": false,
806
+ "normalized": false,
807
+ "rstrip": false,
808
+ "single_word": false,
809
+ "special": true
810
+ },
811
+ "128101": {
812
+ "content": "<|reserved_special_token_93|>",
813
+ "lstrip": false,
814
+ "normalized": false,
815
+ "rstrip": false,
816
+ "single_word": false,
817
+ "special": true
818
+ },
819
+ "128102": {
820
+ "content": "<|reserved_special_token_94|>",
821
+ "lstrip": false,
822
+ "normalized": false,
823
+ "rstrip": false,
824
+ "single_word": false,
825
+ "special": true
826
+ },
827
+ "128103": {
828
+ "content": "<|reserved_special_token_95|>",
829
+ "lstrip": false,
830
+ "normalized": false,
831
+ "rstrip": false,
832
+ "single_word": false,
833
+ "special": true
834
+ },
835
+ "128104": {
836
+ "content": "<|reserved_special_token_96|>",
837
+ "lstrip": false,
838
+ "normalized": false,
839
+ "rstrip": false,
840
+ "single_word": false,
841
+ "special": true
842
+ },
843
+ "128105": {
844
+ "content": "<|reserved_special_token_97|>",
845
+ "lstrip": false,
846
+ "normalized": false,
847
+ "rstrip": false,
848
+ "single_word": false,
849
+ "special": true
850
+ },
851
+ "128106": {
852
+ "content": "<|reserved_special_token_98|>",
853
+ "lstrip": false,
854
+ "normalized": false,
855
+ "rstrip": false,
856
+ "single_word": false,
857
+ "special": true
858
+ },
859
+ "128107": {
860
+ "content": "<|reserved_special_token_99|>",
861
+ "lstrip": false,
862
+ "normalized": false,
863
+ "rstrip": false,
864
+ "single_word": false,
865
+ "special": true
866
+ },
867
+ "128108": {
868
+ "content": "<|reserved_special_token_100|>",
869
+ "lstrip": false,
870
+ "normalized": false,
871
+ "rstrip": false,
872
+ "single_word": false,
873
+ "special": true
874
+ },
875
+ "128109": {
876
+ "content": "<|reserved_special_token_101|>",
877
+ "lstrip": false,
878
+ "normalized": false,
879
+ "rstrip": false,
880
+ "single_word": false,
881
+ "special": true
882
+ },
883
+ "128110": {
884
+ "content": "<|reserved_special_token_102|>",
885
+ "lstrip": false,
886
+ "normalized": false,
887
+ "rstrip": false,
888
+ "single_word": false,
889
+ "special": true
890
+ },
891
+ "128111": {
892
+ "content": "<|reserved_special_token_103|>",
893
+ "lstrip": false,
894
+ "normalized": false,
895
+ "rstrip": false,
896
+ "single_word": false,
897
+ "special": true
898
+ },
899
+ "128112": {
900
+ "content": "<|reserved_special_token_104|>",
901
+ "lstrip": false,
902
+ "normalized": false,
903
+ "rstrip": false,
904
+ "single_word": false,
905
+ "special": true
906
+ },
907
+ "128113": {
908
+ "content": "<|reserved_special_token_105|>",
909
+ "lstrip": false,
910
+ "normalized": false,
911
+ "rstrip": false,
912
+ "single_word": false,
913
+ "special": true
914
+ },
915
+ "128114": {
916
+ "content": "<|reserved_special_token_106|>",
917
+ "lstrip": false,
918
+ "normalized": false,
919
+ "rstrip": false,
920
+ "single_word": false,
921
+ "special": true
922
+ },
923
+ "128115": {
924
+ "content": "<|reserved_special_token_107|>",
925
+ "lstrip": false,
926
+ "normalized": false,
927
+ "rstrip": false,
928
+ "single_word": false,
929
+ "special": true
930
+ },
931
+ "128116": {
932
+ "content": "<|reserved_special_token_108|>",
933
+ "lstrip": false,
934
+ "normalized": false,
935
+ "rstrip": false,
936
+ "single_word": false,
937
+ "special": true
938
+ },
939
+ "128117": {
940
+ "content": "<|reserved_special_token_109|>",
941
+ "lstrip": false,
942
+ "normalized": false,
943
+ "rstrip": false,
944
+ "single_word": false,
945
+ "special": true
946
+ },
947
+ "128118": {
948
+ "content": "<|reserved_special_token_110|>",
949
+ "lstrip": false,
950
+ "normalized": false,
951
+ "rstrip": false,
952
+ "single_word": false,
953
+ "special": true
954
+ },
955
+ "128119": {
956
+ "content": "<|reserved_special_token_111|>",
957
+ "lstrip": false,
958
+ "normalized": false,
959
+ "rstrip": false,
960
+ "single_word": false,
961
+ "special": true
962
+ },
963
+ "128120": {
964
+ "content": "<|reserved_special_token_112|>",
965
+ "lstrip": false,
966
+ "normalized": false,
967
+ "rstrip": false,
968
+ "single_word": false,
969
+ "special": true
970
+ },
971
+ "128121": {
972
+ "content": "<|reserved_special_token_113|>",
973
+ "lstrip": false,
974
+ "normalized": false,
975
+ "rstrip": false,
976
+ "single_word": false,
977
+ "special": true
978
+ },
979
+ "128122": {
980
+ "content": "<|reserved_special_token_114|>",
981
+ "lstrip": false,
982
+ "normalized": false,
983
+ "rstrip": false,
984
+ "single_word": false,
985
+ "special": true
986
+ },
987
+ "128123": {
988
+ "content": "<|reserved_special_token_115|>",
989
+ "lstrip": false,
990
+ "normalized": false,
991
+ "rstrip": false,
992
+ "single_word": false,
993
+ "special": true
994
+ },
995
+ "128124": {
996
+ "content": "<|reserved_special_token_116|>",
997
+ "lstrip": false,
998
+ "normalized": false,
999
+ "rstrip": false,
1000
+ "single_word": false,
1001
+ "special": true
1002
+ },
1003
+ "128125": {
1004
+ "content": "<|reserved_special_token_117|>",
1005
+ "lstrip": false,
1006
+ "normalized": false,
1007
+ "rstrip": false,
1008
+ "single_word": false,
1009
+ "special": true
1010
+ },
1011
+ "128126": {
1012
+ "content": "<|reserved_special_token_118|>",
1013
+ "lstrip": false,
1014
+ "normalized": false,
1015
+ "rstrip": false,
1016
+ "single_word": false,
1017
+ "special": true
1018
+ },
1019
+ "128127": {
1020
+ "content": "<|reserved_special_token_119|>",
1021
+ "lstrip": false,
1022
+ "normalized": false,
1023
+ "rstrip": false,
1024
+ "single_word": false,
1025
+ "special": true
1026
+ },
1027
+ "128128": {
1028
+ "content": "<|reserved_special_token_120|>",
1029
+ "lstrip": false,
1030
+ "normalized": false,
1031
+ "rstrip": false,
1032
+ "single_word": false,
1033
+ "special": true
1034
+ },
1035
+ "128129": {
1036
+ "content": "<|reserved_special_token_121|>",
1037
+ "lstrip": false,
1038
+ "normalized": false,
1039
+ "rstrip": false,
1040
+ "single_word": false,
1041
+ "special": true
1042
+ },
1043
+ "128130": {
1044
+ "content": "<|reserved_special_token_122|>",
1045
+ "lstrip": false,
1046
+ "normalized": false,
1047
+ "rstrip": false,
1048
+ "single_word": false,
1049
+ "special": true
1050
+ },
1051
+ "128131": {
1052
+ "content": "<|reserved_special_token_123|>",
1053
+ "lstrip": false,
1054
+ "normalized": false,
1055
+ "rstrip": false,
1056
+ "single_word": false,
1057
+ "special": true
1058
+ },
1059
+ "128132": {
1060
+ "content": "<|reserved_special_token_124|>",
1061
+ "lstrip": false,
1062
+ "normalized": false,
1063
+ "rstrip": false,
1064
+ "single_word": false,
1065
+ "special": true
1066
+ },
1067
+ "128133": {
1068
+ "content": "<|reserved_special_token_125|>",
1069
+ "lstrip": false,
1070
+ "normalized": false,
1071
+ "rstrip": false,
1072
+ "single_word": false,
1073
+ "special": true
1074
+ },
1075
+ "128134": {
1076
+ "content": "<|reserved_special_token_126|>",
1077
+ "lstrip": false,
1078
+ "normalized": false,
1079
+ "rstrip": false,
1080
+ "single_word": false,
1081
+ "special": true
1082
+ },
1083
+ "128135": {
1084
+ "content": "<|reserved_special_token_127|>",
1085
+ "lstrip": false,
1086
+ "normalized": false,
1087
+ "rstrip": false,
1088
+ "single_word": false,
1089
+ "special": true
1090
+ },
1091
+ "128136": {
1092
+ "content": "<|reserved_special_token_128|>",
1093
+ "lstrip": false,
1094
+ "normalized": false,
1095
+ "rstrip": false,
1096
+ "single_word": false,
1097
+ "special": true
1098
+ },
1099
+ "128137": {
1100
+ "content": "<|reserved_special_token_129|>",
1101
+ "lstrip": false,
1102
+ "normalized": false,
1103
+ "rstrip": false,
1104
+ "single_word": false,
1105
+ "special": true
1106
+ },
1107
+ "128138": {
1108
+ "content": "<|reserved_special_token_130|>",
1109
+ "lstrip": false,
1110
+ "normalized": false,
1111
+ "rstrip": false,
1112
+ "single_word": false,
1113
+ "special": true
1114
+ },
1115
+ "128139": {
1116
+ "content": "<|reserved_special_token_131|>",
1117
+ "lstrip": false,
1118
+ "normalized": false,
1119
+ "rstrip": false,
1120
+ "single_word": false,
1121
+ "special": true
1122
+ },
1123
+ "128140": {
1124
+ "content": "<|reserved_special_token_132|>",
1125
+ "lstrip": false,
1126
+ "normalized": false,
1127
+ "rstrip": false,
1128
+ "single_word": false,
1129
+ "special": true
1130
+ },
1131
+ "128141": {
1132
+ "content": "<|reserved_special_token_133|>",
1133
+ "lstrip": false,
1134
+ "normalized": false,
1135
+ "rstrip": false,
1136
+ "single_word": false,
1137
+ "special": true
1138
+ },
1139
+ "128142": {
1140
+ "content": "<|reserved_special_token_134|>",
1141
+ "lstrip": false,
1142
+ "normalized": false,
1143
+ "rstrip": false,
1144
+ "single_word": false,
1145
+ "special": true
1146
+ },
1147
+ "128143": {
1148
+ "content": "<|reserved_special_token_135|>",
1149
+ "lstrip": false,
1150
+ "normalized": false,
1151
+ "rstrip": false,
1152
+ "single_word": false,
1153
+ "special": true
1154
+ },
1155
+ "128144": {
1156
+ "content": "<|reserved_special_token_136|>",
1157
+ "lstrip": false,
1158
+ "normalized": false,
1159
+ "rstrip": false,
1160
+ "single_word": false,
1161
+ "special": true
1162
+ },
1163
+ "128145": {
1164
+ "content": "<|reserved_special_token_137|>",
1165
+ "lstrip": false,
1166
+ "normalized": false,
1167
+ "rstrip": false,
1168
+ "single_word": false,
1169
+ "special": true
1170
+ },
1171
+ "128146": {
1172
+ "content": "<|reserved_special_token_138|>",
1173
+ "lstrip": false,
1174
+ "normalized": false,
1175
+ "rstrip": false,
1176
+ "single_word": false,
1177
+ "special": true
1178
+ },
1179
+ "128147": {
1180
+ "content": "<|reserved_special_token_139|>",
1181
+ "lstrip": false,
1182
+ "normalized": false,
1183
+ "rstrip": false,
1184
+ "single_word": false,
1185
+ "special": true
1186
+ },
1187
+ "128148": {
1188
+ "content": "<|reserved_special_token_140|>",
1189
+ "lstrip": false,
1190
+ "normalized": false,
1191
+ "rstrip": false,
1192
+ "single_word": false,
1193
+ "special": true
1194
+ },
1195
+ "128149": {
1196
+ "content": "<|reserved_special_token_141|>",
1197
+ "lstrip": false,
1198
+ "normalized": false,
1199
+ "rstrip": false,
1200
+ "single_word": false,
1201
+ "special": true
1202
+ },
1203
+ "128150": {
1204
+ "content": "<|reserved_special_token_142|>",
1205
+ "lstrip": false,
1206
+ "normalized": false,
1207
+ "rstrip": false,
1208
+ "single_word": false,
1209
+ "special": true
1210
+ },
1211
+ "128151": {
1212
+ "content": "<|reserved_special_token_143|>",
1213
+ "lstrip": false,
1214
+ "normalized": false,
1215
+ "rstrip": false,
1216
+ "single_word": false,
1217
+ "special": true
1218
+ },
1219
+ "128152": {
1220
+ "content": "<|reserved_special_token_144|>",
1221
+ "lstrip": false,
1222
+ "normalized": false,
1223
+ "rstrip": false,
1224
+ "single_word": false,
1225
+ "special": true
1226
+ },
1227
+ "128153": {
1228
+ "content": "<|reserved_special_token_145|>",
1229
+ "lstrip": false,
1230
+ "normalized": false,
1231
+ "rstrip": false,
1232
+ "single_word": false,
1233
+ "special": true
1234
+ },
1235
+ "128154": {
1236
+ "content": "<|reserved_special_token_146|>",
1237
+ "lstrip": false,
1238
+ "normalized": false,
1239
+ "rstrip": false,
1240
+ "single_word": false,
1241
+ "special": true
1242
+ },
1243
+ "128155": {
1244
+ "content": "<|reserved_special_token_147|>",
1245
+ "lstrip": false,
1246
+ "normalized": false,
1247
+ "rstrip": false,
1248
+ "single_word": false,
1249
+ "special": true
1250
+ },
1251
+ "128156": {
1252
+ "content": "<|reserved_special_token_148|>",
1253
+ "lstrip": false,
1254
+ "normalized": false,
1255
+ "rstrip": false,
1256
+ "single_word": false,
1257
+ "special": true
1258
+ },
1259
+ "128157": {
1260
+ "content": "<|reserved_special_token_149|>",
1261
+ "lstrip": false,
1262
+ "normalized": false,
1263
+ "rstrip": false,
1264
+ "single_word": false,
1265
+ "special": true
1266
+ },
1267
+ "128158": {
1268
+ "content": "<|reserved_special_token_150|>",
1269
+ "lstrip": false,
1270
+ "normalized": false,
1271
+ "rstrip": false,
1272
+ "single_word": false,
1273
+ "special": true
1274
+ },
1275
+ "128159": {
1276
+ "content": "<|reserved_special_token_151|>",
1277
+ "lstrip": false,
1278
+ "normalized": false,
1279
+ "rstrip": false,
1280
+ "single_word": false,
1281
+ "special": true
1282
+ },
1283
+ "128160": {
1284
+ "content": "<|reserved_special_token_152|>",
1285
+ "lstrip": false,
1286
+ "normalized": false,
1287
+ "rstrip": false,
1288
+ "single_word": false,
1289
+ "special": true
1290
+ },
1291
+ "128161": {
1292
+ "content": "<|reserved_special_token_153|>",
1293
+ "lstrip": false,
1294
+ "normalized": false,
1295
+ "rstrip": false,
1296
+ "single_word": false,
1297
+ "special": true
1298
+ },
1299
+ "128162": {
1300
+ "content": "<|reserved_special_token_154|>",
1301
+ "lstrip": false,
1302
+ "normalized": false,
1303
+ "rstrip": false,
1304
+ "single_word": false,
1305
+ "special": true
1306
+ },
1307
+ "128163": {
1308
+ "content": "<|reserved_special_token_155|>",
1309
+ "lstrip": false,
1310
+ "normalized": false,
1311
+ "rstrip": false,
1312
+ "single_word": false,
1313
+ "special": true
1314
+ },
1315
+ "128164": {
1316
+ "content": "<|reserved_special_token_156|>",
1317
+ "lstrip": false,
1318
+ "normalized": false,
1319
+ "rstrip": false,
1320
+ "single_word": false,
1321
+ "special": true
1322
+ },
1323
+ "128165": {
1324
+ "content": "<|reserved_special_token_157|>",
1325
+ "lstrip": false,
1326
+ "normalized": false,
1327
+ "rstrip": false,
1328
+ "single_word": false,
1329
+ "special": true
1330
+ },
1331
+ "128166": {
1332
+ "content": "<|reserved_special_token_158|>",
1333
+ "lstrip": false,
1334
+ "normalized": false,
1335
+ "rstrip": false,
1336
+ "single_word": false,
1337
+ "special": true
1338
+ },
1339
+ "128167": {
1340
+ "content": "<|reserved_special_token_159|>",
1341
+ "lstrip": false,
1342
+ "normalized": false,
1343
+ "rstrip": false,
1344
+ "single_word": false,
1345
+ "special": true
1346
+ },
1347
+ "128168": {
1348
+ "content": "<|reserved_special_token_160|>",
1349
+ "lstrip": false,
1350
+ "normalized": false,
1351
+ "rstrip": false,
1352
+ "single_word": false,
1353
+ "special": true
1354
+ },
1355
+ "128169": {
1356
+ "content": "<|reserved_special_token_161|>",
1357
+ "lstrip": false,
1358
+ "normalized": false,
1359
+ "rstrip": false,
1360
+ "single_word": false,
1361
+ "special": true
1362
+ },
1363
+ "128170": {
1364
+ "content": "<|reserved_special_token_162|>",
1365
+ "lstrip": false,
1366
+ "normalized": false,
1367
+ "rstrip": false,
1368
+ "single_word": false,
1369
+ "special": true
1370
+ },
1371
+ "128171": {
1372
+ "content": "<|reserved_special_token_163|>",
1373
+ "lstrip": false,
1374
+ "normalized": false,
1375
+ "rstrip": false,
1376
+ "single_word": false,
1377
+ "special": true
1378
+ },
1379
+ "128172": {
1380
+ "content": "<|reserved_special_token_164|>",
1381
+ "lstrip": false,
1382
+ "normalized": false,
1383
+ "rstrip": false,
1384
+ "single_word": false,
1385
+ "special": true
1386
+ },
1387
+ "128173": {
1388
+ "content": "<|reserved_special_token_165|>",
1389
+ "lstrip": false,
1390
+ "normalized": false,
1391
+ "rstrip": false,
1392
+ "single_word": false,
1393
+ "special": true
1394
+ },
1395
+ "128174": {
1396
+ "content": "<|reserved_special_token_166|>",
1397
+ "lstrip": false,
1398
+ "normalized": false,
1399
+ "rstrip": false,
1400
+ "single_word": false,
1401
+ "special": true
1402
+ },
1403
+ "128175": {
1404
+ "content": "<|reserved_special_token_167|>",
1405
+ "lstrip": false,
1406
+ "normalized": false,
1407
+ "rstrip": false,
1408
+ "single_word": false,
1409
+ "special": true
1410
+ },
1411
+ "128176": {
1412
+ "content": "<|reserved_special_token_168|>",
1413
+ "lstrip": false,
1414
+ "normalized": false,
1415
+ "rstrip": false,
1416
+ "single_word": false,
1417
+ "special": true
1418
+ },
1419
+ "128177": {
1420
+ "content": "<|reserved_special_token_169|>",
1421
+ "lstrip": false,
1422
+ "normalized": false,
1423
+ "rstrip": false,
1424
+ "single_word": false,
1425
+ "special": true
1426
+ },
1427
+ "128178": {
1428
+ "content": "<|reserved_special_token_170|>",
1429
+ "lstrip": false,
1430
+ "normalized": false,
1431
+ "rstrip": false,
1432
+ "single_word": false,
1433
+ "special": true
1434
+ },
1435
+ "128179": {
1436
+ "content": "<|reserved_special_token_171|>",
1437
+ "lstrip": false,
1438
+ "normalized": false,
1439
+ "rstrip": false,
1440
+ "single_word": false,
1441
+ "special": true
1442
+ },
1443
+ "128180": {
1444
+ "content": "<|reserved_special_token_172|>",
1445
+ "lstrip": false,
1446
+ "normalized": false,
1447
+ "rstrip": false,
1448
+ "single_word": false,
1449
+ "special": true
1450
+ },
1451
+ "128181": {
1452
+ "content": "<|reserved_special_token_173|>",
1453
+ "lstrip": false,
1454
+ "normalized": false,
1455
+ "rstrip": false,
1456
+ "single_word": false,
1457
+ "special": true
1458
+ },
1459
+ "128182": {
1460
+ "content": "<|reserved_special_token_174|>",
1461
+ "lstrip": false,
1462
+ "normalized": false,
1463
+ "rstrip": false,
1464
+ "single_word": false,
1465
+ "special": true
1466
+ },
1467
+ "128183": {
1468
+ "content": "<|reserved_special_token_175|>",
1469
+ "lstrip": false,
1470
+ "normalized": false,
1471
+ "rstrip": false,
1472
+ "single_word": false,
1473
+ "special": true
1474
+ },
1475
+ "128184": {
1476
+ "content": "<|reserved_special_token_176|>",
1477
+ "lstrip": false,
1478
+ "normalized": false,
1479
+ "rstrip": false,
1480
+ "single_word": false,
1481
+ "special": true
1482
+ },
1483
+ "128185": {
1484
+ "content": "<|reserved_special_token_177|>",
1485
+ "lstrip": false,
1486
+ "normalized": false,
1487
+ "rstrip": false,
1488
+ "single_word": false,
1489
+ "special": true
1490
+ },
1491
+ "128186": {
1492
+ "content": "<|reserved_special_token_178|>",
1493
+ "lstrip": false,
1494
+ "normalized": false,
1495
+ "rstrip": false,
1496
+ "single_word": false,
1497
+ "special": true
1498
+ },
1499
+ "128187": {
1500
+ "content": "<|reserved_special_token_179|>",
1501
+ "lstrip": false,
1502
+ "normalized": false,
1503
+ "rstrip": false,
1504
+ "single_word": false,
1505
+ "special": true
1506
+ },
1507
+ "128188": {
1508
+ "content": "<|reserved_special_token_180|>",
1509
+ "lstrip": false,
1510
+ "normalized": false,
1511
+ "rstrip": false,
1512
+ "single_word": false,
1513
+ "special": true
1514
+ },
1515
+ "128189": {
1516
+ "content": "<|reserved_special_token_181|>",
1517
+ "lstrip": false,
1518
+ "normalized": false,
1519
+ "rstrip": false,
1520
+ "single_word": false,
1521
+ "special": true
1522
+ },
1523
+ "128190": {
1524
+ "content": "<|reserved_special_token_182|>",
1525
+ "lstrip": false,
1526
+ "normalized": false,
1527
+ "rstrip": false,
1528
+ "single_word": false,
1529
+ "special": true
1530
+ },
1531
+ "128191": {
1532
+ "content": "<|reserved_special_token_183|>",
1533
+ "lstrip": false,
1534
+ "normalized": false,
1535
+ "rstrip": false,
1536
+ "single_word": false,
1537
+ "special": true
1538
+ },
1539
+ "128192": {
1540
+ "content": "<|reserved_special_token_184|>",
1541
+ "lstrip": false,
1542
+ "normalized": false,
1543
+ "rstrip": false,
1544
+ "single_word": false,
1545
+ "special": true
1546
+ },
1547
+ "128193": {
1548
+ "content": "<|reserved_special_token_185|>",
1549
+ "lstrip": false,
1550
+ "normalized": false,
1551
+ "rstrip": false,
1552
+ "single_word": false,
1553
+ "special": true
1554
+ },
1555
+ "128194": {
1556
+ "content": "<|reserved_special_token_186|>",
1557
+ "lstrip": false,
1558
+ "normalized": false,
1559
+ "rstrip": false,
1560
+ "single_word": false,
1561
+ "special": true
1562
+ },
1563
+ "128195": {
1564
+ "content": "<|reserved_special_token_187|>",
1565
+ "lstrip": false,
1566
+ "normalized": false,
1567
+ "rstrip": false,
1568
+ "single_word": false,
1569
+ "special": true
1570
+ },
1571
+ "128196": {
1572
+ "content": "<|reserved_special_token_188|>",
1573
+ "lstrip": false,
1574
+ "normalized": false,
1575
+ "rstrip": false,
1576
+ "single_word": false,
1577
+ "special": true
1578
+ },
1579
+ "128197": {
1580
+ "content": "<|reserved_special_token_189|>",
1581
+ "lstrip": false,
1582
+ "normalized": false,
1583
+ "rstrip": false,
1584
+ "single_word": false,
1585
+ "special": true
1586
+ },
1587
+ "128198": {
1588
+ "content": "<|reserved_special_token_190|>",
1589
+ "lstrip": false,
1590
+ "normalized": false,
1591
+ "rstrip": false,
1592
+ "single_word": false,
1593
+ "special": true
1594
+ },
1595
+ "128199": {
1596
+ "content": "<|reserved_special_token_191|>",
1597
+ "lstrip": false,
1598
+ "normalized": false,
1599
+ "rstrip": false,
1600
+ "single_word": false,
1601
+ "special": true
1602
+ },
1603
+ "128200": {
1604
+ "content": "<|reserved_special_token_192|>",
1605
+ "lstrip": false,
1606
+ "normalized": false,
1607
+ "rstrip": false,
1608
+ "single_word": false,
1609
+ "special": true
1610
+ },
1611
+ "128201": {
1612
+ "content": "<|reserved_special_token_193|>",
1613
+ "lstrip": false,
1614
+ "normalized": false,
1615
+ "rstrip": false,
1616
+ "single_word": false,
1617
+ "special": true
1618
+ },
1619
+ "128202": {
1620
+ "content": "<|reserved_special_token_194|>",
1621
+ "lstrip": false,
1622
+ "normalized": false,
1623
+ "rstrip": false,
1624
+ "single_word": false,
1625
+ "special": true
1626
+ },
1627
+ "128203": {
1628
+ "content": "<|reserved_special_token_195|>",
1629
+ "lstrip": false,
1630
+ "normalized": false,
1631
+ "rstrip": false,
1632
+ "single_word": false,
1633
+ "special": true
1634
+ },
1635
+ "128204": {
1636
+ "content": "<|reserved_special_token_196|>",
1637
+ "lstrip": false,
1638
+ "normalized": false,
1639
+ "rstrip": false,
1640
+ "single_word": false,
1641
+ "special": true
1642
+ },
1643
+ "128205": {
1644
+ "content": "<|reserved_special_token_197|>",
1645
+ "lstrip": false,
1646
+ "normalized": false,
1647
+ "rstrip": false,
1648
+ "single_word": false,
1649
+ "special": true
1650
+ },
1651
+ "128206": {
1652
+ "content": "<|reserved_special_token_198|>",
1653
+ "lstrip": false,
1654
+ "normalized": false,
1655
+ "rstrip": false,
1656
+ "single_word": false,
1657
+ "special": true
1658
+ },
1659
+ "128207": {
1660
+ "content": "<|reserved_special_token_199|>",
1661
+ "lstrip": false,
1662
+ "normalized": false,
1663
+ "rstrip": false,
1664
+ "single_word": false,
1665
+ "special": true
1666
+ },
1667
+ "128208": {
1668
+ "content": "<|reserved_special_token_200|>",
1669
+ "lstrip": false,
1670
+ "normalized": false,
1671
+ "rstrip": false,
1672
+ "single_word": false,
1673
+ "special": true
1674
+ },
1675
+ "128209": {
1676
+ "content": "<|reserved_special_token_201|>",
1677
+ "lstrip": false,
1678
+ "normalized": false,
1679
+ "rstrip": false,
1680
+ "single_word": false,
1681
+ "special": true
1682
+ },
1683
+ "128210": {
1684
+ "content": "<|reserved_special_token_202|>",
1685
+ "lstrip": false,
1686
+ "normalized": false,
1687
+ "rstrip": false,
1688
+ "single_word": false,
1689
+ "special": true
1690
+ },
1691
+ "128211": {
1692
+ "content": "<|reserved_special_token_203|>",
1693
+ "lstrip": false,
1694
+ "normalized": false,
1695
+ "rstrip": false,
1696
+ "single_word": false,
1697
+ "special": true
1698
+ },
1699
+ "128212": {
1700
+ "content": "<|reserved_special_token_204|>",
1701
+ "lstrip": false,
1702
+ "normalized": false,
1703
+ "rstrip": false,
1704
+ "single_word": false,
1705
+ "special": true
1706
+ },
1707
+ "128213": {
1708
+ "content": "<|reserved_special_token_205|>",
1709
+ "lstrip": false,
1710
+ "normalized": false,
1711
+ "rstrip": false,
1712
+ "single_word": false,
1713
+ "special": true
1714
+ },
1715
+ "128214": {
1716
+ "content": "<|reserved_special_token_206|>",
1717
+ "lstrip": false,
1718
+ "normalized": false,
1719
+ "rstrip": false,
1720
+ "single_word": false,
1721
+ "special": true
1722
+ },
1723
+ "128215": {
1724
+ "content": "<|reserved_special_token_207|>",
1725
+ "lstrip": false,
1726
+ "normalized": false,
1727
+ "rstrip": false,
1728
+ "single_word": false,
1729
+ "special": true
1730
+ },
1731
+ "128216": {
1732
+ "content": "<|reserved_special_token_208|>",
1733
+ "lstrip": false,
1734
+ "normalized": false,
1735
+ "rstrip": false,
1736
+ "single_word": false,
1737
+ "special": true
1738
+ },
1739
+ "128217": {
1740
+ "content": "<|reserved_special_token_209|>",
1741
+ "lstrip": false,
1742
+ "normalized": false,
1743
+ "rstrip": false,
1744
+ "single_word": false,
1745
+ "special": true
1746
+ },
1747
+ "128218": {
1748
+ "content": "<|reserved_special_token_210|>",
1749
+ "lstrip": false,
1750
+ "normalized": false,
1751
+ "rstrip": false,
1752
+ "single_word": false,
1753
+ "special": true
1754
+ },
1755
+ "128219": {
1756
+ "content": "<|reserved_special_token_211|>",
1757
+ "lstrip": false,
1758
+ "normalized": false,
1759
+ "rstrip": false,
1760
+ "single_word": false,
1761
+ "special": true
1762
+ },
1763
+ "128220": {
1764
+ "content": "<|reserved_special_token_212|>",
1765
+ "lstrip": false,
1766
+ "normalized": false,
1767
+ "rstrip": false,
1768
+ "single_word": false,
1769
+ "special": true
1770
+ },
1771
+ "128221": {
1772
+ "content": "<|reserved_special_token_213|>",
1773
+ "lstrip": false,
1774
+ "normalized": false,
1775
+ "rstrip": false,
1776
+ "single_word": false,
1777
+ "special": true
1778
+ },
1779
+ "128222": {
1780
+ "content": "<|reserved_special_token_214|>",
1781
+ "lstrip": false,
1782
+ "normalized": false,
1783
+ "rstrip": false,
1784
+ "single_word": false,
1785
+ "special": true
1786
+ },
1787
+ "128223": {
1788
+ "content": "<|reserved_special_token_215|>",
1789
+ "lstrip": false,
1790
+ "normalized": false,
1791
+ "rstrip": false,
1792
+ "single_word": false,
1793
+ "special": true
1794
+ },
1795
+ "128224": {
1796
+ "content": "<|reserved_special_token_216|>",
1797
+ "lstrip": false,
1798
+ "normalized": false,
1799
+ "rstrip": false,
1800
+ "single_word": false,
1801
+ "special": true
1802
+ },
1803
+ "128225": {
1804
+ "content": "<|reserved_special_token_217|>",
1805
+ "lstrip": false,
1806
+ "normalized": false,
1807
+ "rstrip": false,
1808
+ "single_word": false,
1809
+ "special": true
1810
+ },
1811
+ "128226": {
1812
+ "content": "<|reserved_special_token_218|>",
1813
+ "lstrip": false,
1814
+ "normalized": false,
1815
+ "rstrip": false,
1816
+ "single_word": false,
1817
+ "special": true
1818
+ },
1819
+ "128227": {
1820
+ "content": "<|reserved_special_token_219|>",
1821
+ "lstrip": false,
1822
+ "normalized": false,
1823
+ "rstrip": false,
1824
+ "single_word": false,
1825
+ "special": true
1826
+ },
1827
+ "128228": {
1828
+ "content": "<|reserved_special_token_220|>",
1829
+ "lstrip": false,
1830
+ "normalized": false,
1831
+ "rstrip": false,
1832
+ "single_word": false,
1833
+ "special": true
1834
+ },
1835
+ "128229": {
1836
+ "content": "<|reserved_special_token_221|>",
1837
+ "lstrip": false,
1838
+ "normalized": false,
1839
+ "rstrip": false,
1840
+ "single_word": false,
1841
+ "special": true
1842
+ },
1843
+ "128230": {
1844
+ "content": "<|reserved_special_token_222|>",
1845
+ "lstrip": false,
1846
+ "normalized": false,
1847
+ "rstrip": false,
1848
+ "single_word": false,
1849
+ "special": true
1850
+ },
1851
+ "128231": {
1852
+ "content": "<|reserved_special_token_223|>",
1853
+ "lstrip": false,
1854
+ "normalized": false,
1855
+ "rstrip": false,
1856
+ "single_word": false,
1857
+ "special": true
1858
+ },
1859
+ "128232": {
1860
+ "content": "<|reserved_special_token_224|>",
1861
+ "lstrip": false,
1862
+ "normalized": false,
1863
+ "rstrip": false,
1864
+ "single_word": false,
1865
+ "special": true
1866
+ },
1867
+ "128233": {
1868
+ "content": "<|reserved_special_token_225|>",
1869
+ "lstrip": false,
1870
+ "normalized": false,
1871
+ "rstrip": false,
1872
+ "single_word": false,
1873
+ "special": true
1874
+ },
1875
+ "128234": {
1876
+ "content": "<|reserved_special_token_226|>",
1877
+ "lstrip": false,
1878
+ "normalized": false,
1879
+ "rstrip": false,
1880
+ "single_word": false,
1881
+ "special": true
1882
+ },
1883
+ "128235": {
1884
+ "content": "<|reserved_special_token_227|>",
1885
+ "lstrip": false,
1886
+ "normalized": false,
1887
+ "rstrip": false,
1888
+ "single_word": false,
1889
+ "special": true
1890
+ },
1891
+ "128236": {
1892
+ "content": "<|reserved_special_token_228|>",
1893
+ "lstrip": false,
1894
+ "normalized": false,
1895
+ "rstrip": false,
1896
+ "single_word": false,
1897
+ "special": true
1898
+ },
1899
+ "128237": {
1900
+ "content": "<|reserved_special_token_229|>",
1901
+ "lstrip": false,
1902
+ "normalized": false,
1903
+ "rstrip": false,
1904
+ "single_word": false,
1905
+ "special": true
1906
+ },
1907
+ "128238": {
1908
+ "content": "<|reserved_special_token_230|>",
1909
+ "lstrip": false,
1910
+ "normalized": false,
1911
+ "rstrip": false,
1912
+ "single_word": false,
1913
+ "special": true
1914
+ },
1915
+ "128239": {
1916
+ "content": "<|reserved_special_token_231|>",
1917
+ "lstrip": false,
1918
+ "normalized": false,
1919
+ "rstrip": false,
1920
+ "single_word": false,
1921
+ "special": true
1922
+ },
1923
+ "128240": {
1924
+ "content": "<|reserved_special_token_232|>",
1925
+ "lstrip": false,
1926
+ "normalized": false,
1927
+ "rstrip": false,
1928
+ "single_word": false,
1929
+ "special": true
1930
+ },
1931
+ "128241": {
1932
+ "content": "<|reserved_special_token_233|>",
1933
+ "lstrip": false,
1934
+ "normalized": false,
1935
+ "rstrip": false,
1936
+ "single_word": false,
1937
+ "special": true
1938
+ },
1939
+ "128242": {
1940
+ "content": "<|reserved_special_token_234|>",
1941
+ "lstrip": false,
1942
+ "normalized": false,
1943
+ "rstrip": false,
1944
+ "single_word": false,
1945
+ "special": true
1946
+ },
1947
+ "128243": {
1948
+ "content": "<|reserved_special_token_235|>",
1949
+ "lstrip": false,
1950
+ "normalized": false,
1951
+ "rstrip": false,
1952
+ "single_word": false,
1953
+ "special": true
1954
+ },
1955
+ "128244": {
1956
+ "content": "<|reserved_special_token_236|>",
1957
+ "lstrip": false,
1958
+ "normalized": false,
1959
+ "rstrip": false,
1960
+ "single_word": false,
1961
+ "special": true
1962
+ },
1963
+ "128245": {
1964
+ "content": "<|reserved_special_token_237|>",
1965
+ "lstrip": false,
1966
+ "normalized": false,
1967
+ "rstrip": false,
1968
+ "single_word": false,
1969
+ "special": true
1970
+ },
1971
+ "128246": {
1972
+ "content": "<|reserved_special_token_238|>",
1973
+ "lstrip": false,
1974
+ "normalized": false,
1975
+ "rstrip": false,
1976
+ "single_word": false,
1977
+ "special": true
1978
+ },
1979
+ "128247": {
1980
+ "content": "<|reserved_special_token_239|>",
1981
+ "lstrip": false,
1982
+ "normalized": false,
1983
+ "rstrip": false,
1984
+ "single_word": false,
1985
+ "special": true
1986
+ },
1987
+ "128248": {
1988
+ "content": "<|reserved_special_token_240|>",
1989
+ "lstrip": false,
1990
+ "normalized": false,
1991
+ "rstrip": false,
1992
+ "single_word": false,
1993
+ "special": true
1994
+ },
1995
+ "128249": {
1996
+ "content": "<|reserved_special_token_241|>",
1997
+ "lstrip": false,
1998
+ "normalized": false,
1999
+ "rstrip": false,
2000
+ "single_word": false,
2001
+ "special": true
2002
+ },
2003
+ "128250": {
2004
+ "content": "<|reserved_special_token_242|>",
2005
+ "lstrip": false,
2006
+ "normalized": false,
2007
+ "rstrip": false,
2008
+ "single_word": false,
2009
+ "special": true
2010
+ },
2011
+ "128251": {
2012
+ "content": "<|reserved_special_token_243|>",
2013
+ "lstrip": false,
2014
+ "normalized": false,
2015
+ "rstrip": false,
2016
+ "single_word": false,
2017
+ "special": true
2018
+ },
2019
+ "128252": {
2020
+ "content": "<|reserved_special_token_244|>",
2021
+ "lstrip": false,
2022
+ "normalized": false,
2023
+ "rstrip": false,
2024
+ "single_word": false,
2025
+ "special": true
2026
+ },
2027
+ "128253": {
2028
+ "content": "<|reserved_special_token_245|>",
2029
+ "lstrip": false,
2030
+ "normalized": false,
2031
+ "rstrip": false,
2032
+ "single_word": false,
2033
+ "special": true
2034
+ },
2035
+ "128254": {
2036
+ "content": "<|reserved_special_token_246|>",
2037
+ "lstrip": false,
2038
+ "normalized": false,
2039
+ "rstrip": false,
2040
+ "single_word": false,
2041
+ "special": true
2042
+ },
2043
+ "128255": {
2044
+ "content": "<|reserved_special_token_247|>",
2045
+ "lstrip": false,
2046
+ "normalized": false,
2047
+ "rstrip": false,
2048
+ "single_word": false,
2049
+ "special": true
2050
+ },
2051
+ "128256": {
2052
+ "content": "<thinking>",
2053
+ "lstrip": false,
2054
+ "normalized": false,
2055
+ "rstrip": false,
2056
+ "single_word": false,
2057
+ "special": true
2058
+ },
2059
+ "128257": {
2060
+ "content": "</thinking>",
2061
+ "lstrip": false,
2062
+ "normalized": false,
2063
+ "rstrip": false,
2064
+ "single_word": false,
2065
+ "special": true
2066
+ },
2067
+ "128258": {
2068
+ "content": "<output>",
2069
+ "lstrip": false,
2070
+ "normalized": false,
2071
+ "rstrip": false,
2072
+ "single_word": false,
2073
+ "special": true
2074
+ },
2075
+ "128259": {
2076
+ "content": "</output>",
2077
+ "lstrip": false,
2078
+ "normalized": false,
2079
+ "rstrip": false,
2080
+ "single_word": false,
2081
+ "special": true
2082
+ },
2083
+ "128260": {
2084
+ "content": "<reflection>",
2085
+ "lstrip": false,
2086
+ "normalized": false,
2087
+ "rstrip": false,
2088
+ "single_word": false,
2089
+ "special": true
2090
+ },
2091
+ "128261": {
2092
+ "content": "</reflection>",
2093
+ "lstrip": false,
2094
+ "normalized": false,
2095
+ "rstrip": false,
2096
+ "single_word": false,
2097
+ "special": true
2098
+ }
2099
+ },
2100
+ "additional_special_tokens": [
2101
+ "<thinking>",
2102
+ "</thinking>",
2103
+ "<output>",
2104
+ "</output>",
2105
+ "<reflection>",
2106
+ "</reflection>"
2107
+ ],
2108
+ "bos_token": "<|begin_of_text|>",
2109
+ "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n",
2110
+ "clean_up_tokenization_spaces": true,
2111
+ "eos_token": "<|eot_id|>",
2112
+ "model_input_names": [
2113
+ "input_ids",
2114
+ "attention_mask"
2115
+ ],
2116
+ "model_max_length": 131072,
2117
+ "pad_token": "<|eot_id|>",
2118
+ "tokenizer_class": "PreTrainedTokenizerFast"
2119
+ }