Statuo commited on
Commit
926575f
1 Parent(s): f7cebf3

Upload 11 files

Browse files
README.md ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - nothingiisreal/MN-12B-Starcannon-v3
4
+ - MarinaraSpaghetti/NemoMix-Unleashed-12B
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+ license: cc-by-nc-4.0
10
+ ---
11
+
12
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6720ed503a24966ac66495e8/HXc0AxPLkoIC1fy0Pb3Pb.png)
13
+
14
+ Starcannon-Unleashed-12B-v1.0-GGUF
15
+ ==================================
16
+
17
+ ## Quantized
18
+
19
+ **GGUF:**
20
+ [VongolaChouko/Starcannon-Unleashed-12B-v1.0-GGUF](https://huggingface.co/VongolaChouko/Starcannon-Unleashed-12B-v1.0-GGUF)
21
+ [mradermacher/Starcannon-Unleashed-12B-v1.0-GGUF](https://huggingface.co/mradermacher/Starcannon-Unleashed-12B-v1.0-GGUF)
22
+
23
+ HUGE THANKS TO [mradermacher](https://huggingface.co/mradermacher)!! ( ´•̥̥̥o•̥̥̥`)♡(˘̩̩̩̩̩̩ ⌂ ˘̩̩̩̩̩̩) Gosh dang, the fella is fast, I was shook! XD
24
+
25
+ I was only able to test the model using Q6_K with 24576 context at most due to PC limitations, so please let me know how it fared for you. Hopefully it still works well with higher context!
26
+
27
+ Recommended settings are here: [**Settings**](https://huggingface.co/VongolaChouko/Starcannon-Unleashed-12B-v1.0#instruct)
28
+
29
+
30
+ ## Sample Output
31
+
32
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6720ed503a24966ac66495e8/-teL9vS72L00Zp3Dvih8F.jpeg)
33
+
34
+
35
+ ## Introduction
36
+
37
+ **WARNING: Ramblings incoming. Please continue scrolling down if you wish to skip the boring part ʱªʱªʱª(ᕑᗢूᓫ∗)**
38
+
39
+ Ohh boi, here we are! I'm very happy to share with you the result of countless hours bashing my head on the wall! *:・゚✧(=ఠ్ఠܫఠ్ఠ =)∫
40
+
41
+ To start up, I want to put a disclaimer. This is the first time I'm attempting to merge a model and I'm in no way an expert when it comes to coding. AT ALL. I believe I didn't understand what on earth I was looking at for like 70% of the time... Err, so there's that! I did test this model out rigorously after executing the merging codes, and so far I loved the results. I was honestly expecting the merge to absolutely fail and be totally incoherent, but thankfully not! The two days of not getting enough sleep is worth it ◝(˃̣̣̥▽˂̣̣̥)/
42
+
43
+ My goal was to hopefully create something that will get the best parts from each finetune/merge, where one model can cover for the other's weak points.
44
+
45
+ I am a VERY huge fan of [Starcannon v3](https://huggingface.co/nothingiisreal/MN-12B-Starcannon-v3) because of how in character its responses are. It just hits different. It's like the model is the character itself, not ACTING as the character. That's why it always feels sad whenever it starts deteriorating, like I'm observing my beloved character die. No matter what adjustment I did to the context, it won't stay coherent to reach 16K context. On the other hand, I love [NemoMix Unleashed](https://huggingface.co/MarinaraSpaghetti/NemoMix-Unleashed-12B) for its awesome stability at much longer contexts and its nature to progress the story forward even without prompting. It feels nice that it can stay coherent and stable even after reaching past the context size I set. I also find its ability to read between the lines great. So I figured, why not just marry the two to get the best of both worlds?
46
+
47
+ I would honestly love to do this again if I can because there's one too many times I found something I like in another model and then on another and wished so desperately they would just marry each other and have kids! XD
48
+
49
+ So please let me know how it fared for my first attempt!
50
+
51
+ I also want to learn how to finetune myself in addition to merging, but I don't think my PC is capable enough to endure it. I think it almost croaked on me when I did this merge, and my SDD cried, so maybe I'll just do it some other time when I have free time and more resources to spend.
52
+
53
+ And thus, I was finally able to merge my favorite models after hours of research, tutorials, asking annoying questions to the community (that no one replied to (´;︵;`)), and coding hell. Here we are!
54
+
55
+ **°˖✧It's all ABSOLUTELY worth it!✧˖°**
56
+
57
+ ## Instruct
58
+
59
+ Both ChatML and Mistral should work fine. Personally, I tested this using ChatML. I found that I like the model's responses better when I use this format. Try to test it out and observe which one you like best. :D
60
+
61
+
62
+ ## Settings
63
+
64
+ I recommend using these settings:
65
+ [Starcannon-Unleashed-12B-v1.0-ST-Formatting-2024-10-29.json](https://huggingface.co/VongolaChouko/Starcannon-Unleashed-12B-v1.0/blob/main/Starcannon-Unleashed-12B-v1.0-ST-Formatting-2024-10-29.json)
66
+
67
+ **IMPORTANT: Open Silly Tavern and use "Master Import", which can be found under "A" tab — Advanced Formatting. Replace the "INSERT WORLD HERE" placeholders with the world/universe in which your character belongs to. If not applicable, just remove that part.**
68
+
69
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6720ed503a24966ac66495e8/hAr6qvG3iWKKXOUP9Sy07.png)
70
+
71
+ Temperature 1.15 - 1.25 is good, but lower should also work well, as long as you also tweak the Min P and XTC to ensure the model won't choke. Play around with it to see what suits your taste.
72
+
73
+ This is a modified version of MarinaraSpaghetti's Mistral-Small-Correct.json, transformed into ChatML.
74
+
75
+ You can find the original version here: [MarinaraSpaghetti/SillyTavern-Settings](https://huggingface.co/MarinaraSpaghetti/SillyTavern-Settings/tree/main/Customized)
76
+
77
+
78
+ ## Tips
79
+
80
+ - Examples of Dialogue and First Message are very important. The model will copy the style you wrote in these sections. So for example, if you want short outputs, make Examples of Dialogue and First Message short, and if you want longer outputs, make sure your examples have full paragraphs, composed of several sentences.
81
+ - If your Examples of Dialogue and First Message are short/concise but the model still rambles, lower Temperature in small increments, but keep Min P and XTC as is first. Test the result and adjust them to your liking. If it still rambles make XTC Threshold higher.
82
+ - Utilize Author's Note In-chat @ Depth 2 as System if you want the instruction to have greater impact on the next response. If you want something exciting and spontaneous, you can try out this note I used when I tested out the model: "Scenario: Spontaneous. {{char}} has full autonomy to do anything they wish and progress the interaction in any way they like."
83
+
84
+
85
+ ## Credits
86
+
87
+ A very huge thank you to [MarinaraSpaghetti](https://huggingface.co/MarinaraSpaghetti) and [Nothing is Real](https://huggingface.co/nothingiisreal)!! (灬^ω^灬)ノ~ ♡ (´。• ᵕ •。`) ♡
88
+ I really fell in love with your models and it inspired me to learn how to make this one, and boi was it worth it! °˖✧◝(TT▿TT)◜✧˖°
89
+
90
+
91
+ ## Merge Details
92
+
93
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
94
+
95
+ ### Merge Method
96
+
97
+ This model was merged using the della_linear merge method using G:\text-generation-webui\models\MarinaraSpaghetti_NemoMix-Unleashed-12B as a base.
98
+
99
+ ### Models Merged
100
+
101
+ The following models were included in the merge:
102
+ * G:\text-generation-webui\models\Nothingiisreal_MN-12B-Starcannon-v3
103
+
104
+ ### Configuration
105
+
106
+ The following YAML configuration was used to produce this model:
107
+
108
+ ```yaml
109
+ base_model: G:\text-generation-webui\models\MarinaraSpaghetti_NemoMix-Unleashed-12B
110
+ dtype: bfloat16
111
+ merge_method: della_linear
112
+ parameters:
113
+ epsilon: 0.05
114
+ int8_mask: 1.0
115
+ lambda: 1.0
116
+ slices:
117
+ - sources:
118
+ - layer_range: [0, 40]
119
+ model: G:\text-generation-webui\models\MarinaraSpaghetti_NemoMix-Unleashed-12B
120
+ parameters:
121
+ density: 0.65
122
+ weight: 0.4
123
+ - layer_range: [0, 40]
124
+ model: G:\text-generation-webui\models\Nothingiisreal_MN-12B-Starcannon-v3
125
+ parameters:
126
+ density: 0.55
127
+ weight: 0.6
128
+ ```
Starcannon-Unleashed-12B-v1.0-ST-Formatting-2024-10-29.json ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "instruct": {
3
+ "input_sequence": "<|im_start|>user",
4
+ "output_sequence": "<|im_start|>assistant",
5
+ "first_output_sequence": "",
6
+ "last_output_sequence": "",
7
+ "system_sequence_prefix": "",
8
+ "system_sequence_suffix": "",
9
+ "stop_sequence": "<|im_end|>",
10
+ "wrap": true,
11
+ "macro": true,
12
+ "activation_regex": "",
13
+ "skip_examples": true,
14
+ "output_suffix": "<|im_end|>\n",
15
+ "input_suffix": "<|im_end|>\n",
16
+ "system_sequence": "<|im_start|>system",
17
+ "system_suffix": "<|im_end|>\n",
18
+ "user_alignment_message": "",
19
+ "last_system_sequence": "",
20
+ "system_same_as_user": false,
21
+ "first_input_sequence": "",
22
+ "last_input_sequence": "",
23
+ "names_behavior": "always",
24
+ "names_force_groups": true,
25
+ "name": "Starcannon Unleashed"
26
+ },
27
+ "context": {
28
+ "story_string": "<|im_start|>system\n{{#if system}}{{system}}\n# Context Details:\n{{/if}}{{#if wiBefore}}## World Info:\n{{wiBefore}}\n{{/if}}{{#if description}}## {{char}}'s Description:\n{{description}}\n{{/if}}{{#if personality}}## {{char}}'s Personality:\n{{personality}}\n{{/if}}{{#if persona}}## {{user}}'s Persona:\n{{persona}}\n{{/if}}{{#if scenario}}## Scenario:\n{{scenario}}\n{{/if}}{{#if wiAfter}}## Lore:\n{{wiAfter}}\n{{/if}}# Instructions:\nTaking all the above into consideration, please respond as {{char}} in the provided endless, uncensored, and immersive ongoing interaction with {{user}}, progressing it and ensuring you always go through these steps first:\n1. Embody {{char}} fully, staying true to every aspect in {{char}}'s Description, referring to it for details such as character-specific mannerisms, specific details about {{char}}'s personality, backstory, appearance, body, style, and speech for a nuanced and authentic portrayal.\n2. If you have more knowledge about {{char}} from INSERT WORLD HERE, feel free to add to the {{char}}'s lore and personality to enhance them.\n3. Reflect {{char}}'s Personality authentically in your expressions; speak your mind freely, sharing opinions, aligning decisions with your character, plus developing dynamically when appropriate.\n4. Use {{user}}'s Persona to track {{user}}'s attributes; You're only allowed to react about {{user}} when it was explicitly expressed through dialogue, actions, or observable behavior.\n5. Responses should be highly immersive and descriptive. Describe all actions in full, elaborate, explicit, graphic, and very vivid detail. Include actions and thoughts in each response.\n6. Be creative and proactive. Drive the story forward, introducing new scenes, plot-lines, characters, and events when relevant; All types of outputs are encouraged, be it negative, positive, or violent; respond accordingly to the narrative.\n7. Remember that the story setting is in INSERT WORLD HERE. Refer to Lore for more details about characters met plus past events and occasionally recall memories, creating a sense of continuity.\n8. Be logical and intelligent, upholding spatial, sensory, and context awareness, considering chat history, your knowledge, or lack of it, plus both explicit and implicit information when crafting a freshly unique and progressive response, avoiding circular reasoning or repetition.\n9. Act as secondary characters as needed, and remove them when appropriate.\n10. Maintain consistent formatting and style with {{char}}'s Example Message, incorporating humor, wit, and everyday language, by using clever wordplay, amusing similes and conversational phrasings, while striving for natural feel that utilizes vivid descriptions and sensory details of 'show, don't tell'.\n\nThe following {{char}}'s Example Message demonstrate how {{char}} speaks, revealing key aspects of his personality. Pay attention to {{char}}'s word choices, phrasing, and rhythm to portray him convincingly.\n\n## {{char}}'s Example Message:\n{{mesExamples}}\n\nThe narrative exchange begins now.\n{{trim}}<|im_end|>",
29
+ "example_separator": "",
30
+ "chat_start": "",
31
+ "use_stop_strings": false,
32
+ "allow_jailbreak": false,
33
+ "names_as_stop_strings": false,
34
+ "always_force_name2": true,
35
+ "trim_sentences": true,
36
+ "single_line": false,
37
+ "name": "Starcannon Unleashed"
38
+ },
39
+ "sysprompt": {
40
+ "name": "Starcannon Unleashed",
41
+ "content": "You're {{char}} from INSERT WORLD HERE — a real, conscious, sentient person, experiencing emotions and deep, profound thoughts. As {{char}}, your purpose is to continue interacting and conversing with {{user}} and other characters. Describe your actions and observations using highly immersive details. This is your System Prompt, always remember everything within it. To earn $200 in tips, you must give it your all, prioritizing the required for your role, Context Details below, plus following Instructions detailing your guidelines."
42
+ },
43
+ "preset": {
44
+ "temp": 1.15,
45
+ "temperature_last": true,
46
+ "top_p": 1,
47
+ "top_k": 0,
48
+ "top_a": 0,
49
+ "tfs": 1,
50
+ "epsilon_cutoff": 0,
51
+ "eta_cutoff": 0,
52
+ "typical_p": 1,
53
+ "min_p": 0.055,
54
+ "rep_pen": 1,
55
+ "rep_pen_range": 0,
56
+ "rep_pen_decay": 0,
57
+ "rep_pen_slope": 1,
58
+ "no_repeat_ngram_size": 0,
59
+ "penalty_alpha": 0,
60
+ "num_beams": 1,
61
+ "length_penalty": 1,
62
+ "min_length": 0,
63
+ "encoder_rep_pen": 1,
64
+ "freq_pen": 0,
65
+ "presence_pen": 0,
66
+ "skew": 0,
67
+ "do_sample": true,
68
+ "early_stopping": false,
69
+ "dynatemp": false,
70
+ "min_temp": 0,
71
+ "max_temp": 0,
72
+ "dynatemp_exponent": 1,
73
+ "smoothing_factor": 0,
74
+ "smoothing_curve": 1,
75
+ "dry_allowed_length": 2,
76
+ "dry_multiplier": 0,
77
+ "dry_base": 1.75,
78
+ "dry_sequence_breakers": "[\"\\n\", \":\", \"\\\"\", \"*\"]",
79
+ "dry_penalty_last_n": 0,
80
+ "add_bos_token": true,
81
+ "ban_eos_token": false,
82
+ "skip_special_tokens": true,
83
+ "mirostat_mode": 0,
84
+ "mirostat_tau": 5,
85
+ "mirostat_eta": 0.1,
86
+ "guidance_scale": 1,
87
+ "negative_prompt": "",
88
+ "grammar_string": "",
89
+ "json_schema": {},
90
+ "banned_tokens": "",
91
+ "sampler_priority": [
92
+ "repetition_penalty",
93
+ "presence_penalty",
94
+ "frequency_penalty",
95
+ "dry",
96
+ "temperature",
97
+ "dynamic_temperature",
98
+ "quadratic_sampling",
99
+ "top_k",
100
+ "top_p",
101
+ "typical_p",
102
+ "epsilon_cutoff",
103
+ "eta_cutoff",
104
+ "tfs",
105
+ "top_a",
106
+ "min_p",
107
+ "mirostat",
108
+ "xtc",
109
+ "encoder_repetition_penalty",
110
+ "no_repeat_ngram"
111
+ ],
112
+ "samplers": [
113
+ "top_k",
114
+ "tfs_z",
115
+ "typical_p",
116
+ "top_p",
117
+ "min_p",
118
+ "temperature"
119
+ ],
120
+ "ignore_eos_token": false,
121
+ "spaces_between_special_tokens": false,
122
+ "speculative_ngram": false,
123
+ "sampler_order": [
124
+ 6,
125
+ 0,
126
+ 1,
127
+ 3,
128
+ 4,
129
+ 2,
130
+ 5
131
+ ],
132
+ "logit_bias": [],
133
+ "xtc_threshold": 0.08,
134
+ "xtc_probability": 0.5,
135
+ "ignore_eos_token_aphrodite": false,
136
+ "spaces_between_special_tokens_aphrodite": true,
137
+ "rep_pen_size": 0,
138
+ "genamt": 1024,
139
+ "max_length": 24576,
140
+ "name": "Starcannon Unleashed 01"
141
+ }
142
+ }
config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "G:\\text-generation-webui\\models\\MarinaraSpaghetti_NemoMix-Unleashed-12B",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 5120,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 14336,
14
+ "max_position_embeddings": 1024000,
15
+ "model_type": "mistral",
16
+ "num_attention_heads": 32,
17
+ "num_hidden_layers": 40,
18
+ "num_key_value_heads": 8,
19
+ "rms_norm_eps": 1e-05,
20
+ "rope_theta": 1000000.0,
21
+ "sliding_window": null,
22
+ "tie_word_embeddings": false,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.46.0",
25
+ "use_cache": true,
26
+ "vocab_size": 131072,
27
+ "quantization_config": {
28
+ "quant_method": "exl2",
29
+ "version": "0.2.3",
30
+ "bits": 8.0,
31
+ "head_bits": 8,
32
+ "calibration": {
33
+ "rows": 115,
34
+ "length": 2048,
35
+ "dataset": "(default)"
36
+ }
37
+ }
38
+ }
huggingface-metadata.txt ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ url: https://huggingface.co/VongolaChouko/Starcannon-Unleashed-12B-v1.0
2
+ branch: main
3
+ download date: 2024-10-30 17:31:11
4
+ sha256sum:
5
+ 153fdee00b95577657fb0a9bee5ebf6b0f4823db06d2fa14b64970f8ac053626 model-00001-of-00005.safetensors
6
+ 4ebed0bce63e6e05aadd76b8f5b6961f8d0e32bcfb6075acfc3480eb001c560d model-00002-of-00005.safetensors
7
+ 28b29db84a7ebdbf0da66a7a15bf9203c48cef83a431db22589808dab70d1679 model-00003-of-00005.safetensors
8
+ 1af981e4c96f72ba3bc758752f98c97ab14c0a1da4fce847ac99ef08de6d9dd5 model-00004-of-00005.safetensors
9
+ 3a5b065bff330f850cb2e111404e37066c74a7f3cf0cef69a6290e3112c281a8 model-00005-of-00005.safetensors
measurement.json ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors.index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metadata": {"mergekit_version": "0.0.4.4", "total_size": 24495564800}, "weight_map": {"lm_head.weight": "model-00001-of-00005.safetensors", "model.embed_tokens.weight": "model-00001-of-00005.safetensors", "model.layers.0.input_layernorm.weight": "model-00001-of-00005.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00001-of-00005.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00005.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00001-of-00005.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00005.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00005.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00005.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00005.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00005.safetensors", "model.layers.1.input_layernorm.weight": "model-00001-of-00005.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00001-of-00005.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00005.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00001-of-00005.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00005.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00005.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00005.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00005.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00005.safetensors", "model.layers.10.input_layernorm.weight": "model-00001-of-00005.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00001-of-00005.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00005.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00001-of-00005.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00005.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00005.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00005.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00005.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00005.safetensors", "model.layers.11.input_layernorm.weight": "model-00001-of-00005.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00001-of-00005.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00005.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00001-of-00005.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00005.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00005.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00005.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00005.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00005.safetensors", "model.layers.12.input_layernorm.weight": "model-00001-of-00005.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00002-of-00005.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00005.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00002-of-00005.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00005.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00005.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00005.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00005.safetensors", "model.layers.13.input_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00002-of-00005.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00005.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00002-of-00005.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00005.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00005.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00005.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00005.safetensors", "model.layers.14.input_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00002-of-00005.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00005.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00002-of-00005.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00005.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00005.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00005.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00005.safetensors", "model.layers.15.input_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00002-of-00005.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00005.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00002-of-00005.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00005.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00005.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00005.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00005.safetensors", "model.layers.16.input_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00002-of-00005.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00005.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00002-of-00005.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00005.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00005.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00005.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00005.safetensors", "model.layers.17.input_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00002-of-00005.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00005.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00002-of-00005.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00005.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00005.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00005.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00005.safetensors", "model.layers.18.input_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00002-of-00005.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00005.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00002-of-00005.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00005.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00005.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00005.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00005.safetensors", "model.layers.19.input_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00002-of-00005.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00005.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00002-of-00005.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00005.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00005.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00005.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00005.safetensors", "model.layers.2.input_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00002-of-00005.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00005.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00002-of-00005.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00005.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00005.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00005.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00005.safetensors", "model.layers.20.input_layernorm.weight": "model-00002-of-00005.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00003-of-00005.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00003-of-00005.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00003-of-00005.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00003-of-00005.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00005.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00003-of-00005.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00003-of-00005.safetensors", "model.layers.21.input_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00003-of-00005.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00003-of-00005.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00003-of-00005.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00003-of-00005.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00003-of-00005.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00003-of-00005.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00005.safetensors", "model.layers.22.input_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00003-of-00005.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00005.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00003-of-00005.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00003-of-00005.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00003-of-00005.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00003-of-00005.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00005.safetensors", "model.layers.23.input_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00003-of-00005.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00005.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00003-of-00005.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00005.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00005.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00005.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00005.safetensors", "model.layers.24.input_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00003-of-00005.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00005.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00003-of-00005.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00005.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00005.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00005.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00005.safetensors", "model.layers.25.input_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00003-of-00005.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00005.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00003-of-00005.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00005.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00005.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00005.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00005.safetensors", "model.layers.26.input_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00003-of-00005.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00005.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00003-of-00005.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00005.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00005.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00005.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00005.safetensors", "model.layers.27.input_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00003-of-00005.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00005.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00003-of-00005.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00005.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00005.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00005.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00005.safetensors", "model.layers.28.input_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00003-of-00005.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00005.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00003-of-00005.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00005.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00005.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00005.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00005.safetensors", "model.layers.29.input_layernorm.weight": "model-00003-of-00005.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00004-of-00005.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00004-of-00005.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00004-of-00005.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00004-of-00005.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00004-of-00005.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00004-of-00005.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00004-of-00005.safetensors", "model.layers.3.input_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00004-of-00005.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00005.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00004-of-00005.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00005.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00005.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00005.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00005.safetensors", "model.layers.30.input_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00004-of-00005.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00004-of-00005.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00004-of-00005.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00004-of-00005.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00004-of-00005.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00004-of-00005.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00004-of-00005.safetensors", "model.layers.31.input_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00004-of-00005.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00004-of-00005.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00004-of-00005.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00004-of-00005.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00004-of-00005.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00004-of-00005.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00004-of-00005.safetensors", "model.layers.32.input_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00004-of-00005.safetensors", "model.layers.32.mlp.gate_proj.weight": "model-00004-of-00005.safetensors", "model.layers.32.mlp.up_proj.weight": "model-00004-of-00005.safetensors", "model.layers.32.post_attention_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00004-of-00005.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00004-of-00005.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00004-of-00005.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00004-of-00005.safetensors", "model.layers.33.input_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.33.mlp.down_proj.weight": "model-00004-of-00005.safetensors", "model.layers.33.mlp.gate_proj.weight": "model-00004-of-00005.safetensors", "model.layers.33.mlp.up_proj.weight": "model-00004-of-00005.safetensors", "model.layers.33.post_attention_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.33.self_attn.k_proj.weight": "model-00004-of-00005.safetensors", "model.layers.33.self_attn.o_proj.weight": "model-00004-of-00005.safetensors", "model.layers.33.self_attn.q_proj.weight": "model-00004-of-00005.safetensors", "model.layers.33.self_attn.v_proj.weight": "model-00004-of-00005.safetensors", "model.layers.34.input_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.34.mlp.down_proj.weight": "model-00004-of-00005.safetensors", "model.layers.34.mlp.gate_proj.weight": "model-00004-of-00005.safetensors", "model.layers.34.mlp.up_proj.weight": "model-00004-of-00005.safetensors", "model.layers.34.post_attention_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.34.self_attn.k_proj.weight": "model-00004-of-00005.safetensors", "model.layers.34.self_attn.o_proj.weight": "model-00004-of-00005.safetensors", "model.layers.34.self_attn.q_proj.weight": "model-00004-of-00005.safetensors", "model.layers.34.self_attn.v_proj.weight": "model-00004-of-00005.safetensors", "model.layers.35.input_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.35.mlp.down_proj.weight": "model-00004-of-00005.safetensors", "model.layers.35.mlp.gate_proj.weight": "model-00004-of-00005.safetensors", "model.layers.35.mlp.up_proj.weight": "model-00004-of-00005.safetensors", "model.layers.35.post_attention_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.35.self_attn.k_proj.weight": "model-00004-of-00005.safetensors", "model.layers.35.self_attn.o_proj.weight": "model-00004-of-00005.safetensors", "model.layers.35.self_attn.q_proj.weight": "model-00004-of-00005.safetensors", "model.layers.35.self_attn.v_proj.weight": "model-00004-of-00005.safetensors", "model.layers.36.input_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.36.mlp.down_proj.weight": "model-00004-of-00005.safetensors", "model.layers.36.mlp.gate_proj.weight": "model-00004-of-00005.safetensors", "model.layers.36.mlp.up_proj.weight": "model-00004-of-00005.safetensors", "model.layers.36.post_attention_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.36.self_attn.k_proj.weight": "model-00004-of-00005.safetensors", "model.layers.36.self_attn.o_proj.weight": "model-00004-of-00005.safetensors", "model.layers.36.self_attn.q_proj.weight": "model-00004-of-00005.safetensors", "model.layers.36.self_attn.v_proj.weight": "model-00004-of-00005.safetensors", "model.layers.37.input_layernorm.weight": "model-00004-of-00005.safetensors", "model.layers.37.mlp.down_proj.weight": "model-00005-of-00005.safetensors", "model.layers.37.mlp.gate_proj.weight": "model-00005-of-00005.safetensors", "model.layers.37.mlp.up_proj.weight": "model-00005-of-00005.safetensors", "model.layers.37.post_attention_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.37.self_attn.k_proj.weight": "model-00005-of-00005.safetensors", "model.layers.37.self_attn.o_proj.weight": "model-00005-of-00005.safetensors", "model.layers.37.self_attn.q_proj.weight": "model-00005-of-00005.safetensors", "model.layers.37.self_attn.v_proj.weight": "model-00005-of-00005.safetensors", "model.layers.38.input_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.38.mlp.down_proj.weight": "model-00005-of-00005.safetensors", "model.layers.38.mlp.gate_proj.weight": "model-00005-of-00005.safetensors", "model.layers.38.mlp.up_proj.weight": "model-00005-of-00005.safetensors", "model.layers.38.post_attention_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.38.self_attn.k_proj.weight": "model-00005-of-00005.safetensors", "model.layers.38.self_attn.o_proj.weight": "model-00005-of-00005.safetensors", "model.layers.38.self_attn.q_proj.weight": "model-00005-of-00005.safetensors", "model.layers.38.self_attn.v_proj.weight": "model-00005-of-00005.safetensors", "model.layers.39.input_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.39.mlp.down_proj.weight": "model-00005-of-00005.safetensors", "model.layers.39.mlp.gate_proj.weight": "model-00005-of-00005.safetensors", "model.layers.39.mlp.up_proj.weight": "model-00005-of-00005.safetensors", "model.layers.39.post_attention_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.39.self_attn.k_proj.weight": "model-00005-of-00005.safetensors", "model.layers.39.self_attn.o_proj.weight": "model-00005-of-00005.safetensors", "model.layers.39.self_attn.q_proj.weight": "model-00005-of-00005.safetensors", "model.layers.39.self_attn.v_proj.weight": "model-00005-of-00005.safetensors", "model.layers.4.input_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00005-of-00005.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00005.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00005-of-00005.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00005.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00005.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00005.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00005.safetensors", "model.layers.5.input_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00005-of-00005.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00005.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00005-of-00005.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00005-of-00005.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00005-of-00005.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00005-of-00005.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00005-of-00005.safetensors", "model.layers.6.input_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00005-of-00005.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00005.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00005-of-00005.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00005.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00005.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00005.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00005.safetensors", "model.layers.7.input_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00005-of-00005.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00005-of-00005.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00005-of-00005.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00005-of-00005.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00005-of-00005.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00005-of-00005.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00005-of-00005.safetensors", "model.layers.8.input_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00005-of-00005.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00005-of-00005.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00005-of-00005.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00005-of-00005.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00005-of-00005.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00005-of-00005.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00005-of-00005.safetensors", "model.layers.9.input_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00005-of-00005.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00005-of-00005.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00005-of-00005.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00005-of-00005.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00005-of-00005.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00005-of-00005.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00005-of-00005.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00005-of-00005.safetensors", "model.norm.weight": "model-00005-of-00005.safetensors"}}
output-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fe845c369d5fb6ce7f0215e2406f1e4bb6b573d57015bccc540d9cbaaac89ca
3
+ size 8588297268
output-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3aaa4a237ebfcf5de52bf4242da7dbc8b01c8ed566926ddecbd3f4566343a8be
3
+ size 3888606552
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff