update files (#5)
Browse files- update files (5afe6449116b70377c10ff869b82f10d36828c73)
Co-authored-by: haoyanli <[email protected]>
- chat_template.json +3 -0
- config.json +5 -2
- model-00001-of-00003.safetensors +2 -2
- model-00002-of-00003.safetensors +2 -2
- model-00003-of-00003.safetensors +2 -2
- model.safetensors.index.json +47 -119
- preprocessor_config.json +1 -0
chat_template.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- if strftime_now is defined %}\n {%- set date_string = strftime_now(\"%d %b %Y\") %}\n {%- else %}\n {%- set date_string = \"26 Jul 2024\" %}\n {%- endif %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- Find out if there are any images #}\n{% set image_ns = namespace(has_images=false) %} \n{%- for message in messages %}\n {%- for content in message['content'] %}\n {%- if content['type'] == 'image' %}\n {%- set image_ns.has_images = true %}\n {%- endif %}\n {%- endfor %}\n{%- endfor %}\n\n{#- Error out if there are images and system message #}\n{%- if image_ns.has_images and not system_message == \"\" %}\n {{- raise_exception(\"Prompting with images is incompatible with system messages.\") }}\n{%- endif %}\n\n{#- System message if there are no images #}\n{%- if not image_ns.has_images %}\n {{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n {%- if tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n {%- endif %}\n {{- \"Cutting Knowledge Date: December 2023\\n\" }}\n {{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n {%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {%- endif %}\n {{- system_message }}\n {{- \"<|eot_id|>\" }}\n{%- endif %}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n' }}\n {%- if message['content'] is string %}\n {{- message['content'] }}\n {%- else %}\n {%- for content in message['content'] %}\n {%- if content['type'] == 'image' %}\n {{- '<|image|>' }}\n {%- elif content['type'] == 'text' %}\n {{- content['text'] }}\n {%- endif %}\n {%- endfor %}\n {%- endif %}\n {{- '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {{- \"<|eot_id|>\" }}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n"
|
| 3 |
+
}
|
config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "/model_path/Llama-3.2-11B-Vision-Instruct
|
| 3 |
"architectures": [
|
| 4 |
"MllamaForConditionalGeneration"
|
| 5 |
],
|
|
@@ -8,7 +8,10 @@
|
|
| 8 |
"quantization_config": {
|
| 9 |
"activation_scheme": "static",
|
| 10 |
"ignored_layers": [
|
| 11 |
-
"lm_head"
|
|
|
|
|
|
|
|
|
|
| 12 |
],
|
| 13 |
"kv_cache_scheme": "static",
|
| 14 |
"quant_method": "fp8"
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "/model_path/Llama-3.2-11B-Vision-Instruct",
|
| 3 |
"architectures": [
|
| 4 |
"MllamaForConditionalGeneration"
|
| 5 |
],
|
|
|
|
| 8 |
"quantization_config": {
|
| 9 |
"activation_scheme": "static",
|
| 10 |
"ignored_layers": [
|
| 11 |
+
"*lm_head",
|
| 12 |
+
"*cross_attn*",
|
| 13 |
+
"vision_model*",
|
| 14 |
+
"multi_modal_projector"
|
| 15 |
],
|
| 16 |
"kv_cache_scheme": "static",
|
| 17 |
"quant_method": "fp8"
|
model-00001-of-00003.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3a3123a5a86e315ff58920026ef7a912fb13788747d38acbbbec47012e89052b
|
| 3 |
+
size 4984320774
|
model-00002-of-00003.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:da57cc3b84a544a430a742177e2ba98d5870cba85bf0cb7f0c3974f31e73d715
|
| 3 |
+
size 4966480928
|
model-00003-of-00003.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8a6c5ade219cab320a5fef98063d814d9c5005fe4c015f9d75d0c9a6d91bd6b5
|
| 3 |
+
size 3001212912
|
model.safetensors.index.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
-
"total_size":
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"language_model.lm_head.weight": "model-00003-of-00003.safetensors",
|
|
@@ -64,19 +64,19 @@
|
|
| 64 |
"language_model.model.layers.10.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 65 |
"language_model.model.layers.10.mlp.up_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 66 |
"language_model.model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 67 |
-
"language_model.model.layers.10.self_attn.k_proj.input_scale": "model-
|
| 68 |
-
"language_model.model.layers.10.self_attn.k_proj.weight": "model-
|
| 69 |
-
"language_model.model.layers.10.self_attn.k_proj.weight_scale": "model-
|
| 70 |
-
"language_model.model.layers.10.self_attn.kv_scale": "model-
|
| 71 |
"language_model.model.layers.10.self_attn.o_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 72 |
"language_model.model.layers.10.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 73 |
"language_model.model.layers.10.self_attn.o_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 74 |
-
"language_model.model.layers.10.self_attn.q_proj.input_scale": "model-
|
| 75 |
-
"language_model.model.layers.10.self_attn.q_proj.weight": "model-
|
| 76 |
-
"language_model.model.layers.10.self_attn.q_proj.weight_scale": "model-
|
| 77 |
-
"language_model.model.layers.10.self_attn.v_proj.input_scale": "model-
|
| 78 |
-
"language_model.model.layers.10.self_attn.v_proj.weight": "model-
|
| 79 |
-
"language_model.model.layers.10.self_attn.v_proj.weight_scale": "model-
|
| 80 |
"language_model.model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 81 |
"language_model.model.layers.11.mlp.down_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 82 |
"language_model.model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
|
@@ -126,20 +126,11 @@
|
|
| 126 |
"language_model.model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 127 |
"language_model.model.layers.12.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 128 |
"language_model.model.layers.13.cross_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
| 129 |
-
"language_model.model.layers.13.cross_attn.k_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 130 |
"language_model.model.layers.13.cross_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 131 |
-
"language_model.model.layers.13.cross_attn.k_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 132 |
-
"language_model.model.layers.13.cross_attn.kv_scale": "model-00002-of-00003.safetensors",
|
| 133 |
-
"language_model.model.layers.13.cross_attn.o_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 134 |
"language_model.model.layers.13.cross_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 135 |
-
"language_model.model.layers.13.cross_attn.o_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 136 |
"language_model.model.layers.13.cross_attn.q_norm.weight": "model-00002-of-00003.safetensors",
|
| 137 |
-
"language_model.model.layers.13.cross_attn.q_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 138 |
"language_model.model.layers.13.cross_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 139 |
-
"language_model.model.layers.13.cross_attn.q_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 140 |
-
"language_model.model.layers.13.cross_attn.v_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 141 |
"language_model.model.layers.13.cross_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 142 |
-
"language_model.model.layers.13.cross_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 143 |
"language_model.model.layers.13.cross_attn_attn_gate": "model-00002-of-00003.safetensors",
|
| 144 |
"language_model.model.layers.13.cross_attn_mlp_gate": "model-00002-of-00003.safetensors",
|
| 145 |
"language_model.model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
|
@@ -250,20 +241,11 @@
|
|
| 250 |
"language_model.model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 251 |
"language_model.model.layers.17.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 252 |
"language_model.model.layers.18.cross_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
| 253 |
-
"language_model.model.layers.18.cross_attn.k_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 254 |
"language_model.model.layers.18.cross_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 255 |
-
"language_model.model.layers.18.cross_attn.k_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 256 |
-
"language_model.model.layers.18.cross_attn.kv_scale": "model-00002-of-00003.safetensors",
|
| 257 |
-
"language_model.model.layers.18.cross_attn.o_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 258 |
"language_model.model.layers.18.cross_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 259 |
-
"language_model.model.layers.18.cross_attn.o_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 260 |
"language_model.model.layers.18.cross_attn.q_norm.weight": "model-00002-of-00003.safetensors",
|
| 261 |
-
"language_model.model.layers.18.cross_attn.q_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 262 |
"language_model.model.layers.18.cross_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 263 |
-
"language_model.model.layers.18.cross_attn.q_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 264 |
-
"language_model.model.layers.18.cross_attn.v_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 265 |
"language_model.model.layers.18.cross_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 266 |
-
"language_model.model.layers.18.cross_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 267 |
"language_model.model.layers.18.cross_attn_attn_gate": "model-00002-of-00003.safetensors",
|
| 268 |
"language_model.model.layers.18.cross_attn_mlp_gate": "model-00002-of-00003.safetensors",
|
| 269 |
"language_model.model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
|
@@ -398,20 +380,11 @@
|
|
| 398 |
"language_model.model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 399 |
"language_model.model.layers.22.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 400 |
"language_model.model.layers.23.cross_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
| 401 |
-
"language_model.model.layers.23.cross_attn.k_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 402 |
"language_model.model.layers.23.cross_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 403 |
-
"language_model.model.layers.23.cross_attn.k_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 404 |
-
"language_model.model.layers.23.cross_attn.kv_scale": "model-00002-of-00003.safetensors",
|
| 405 |
-
"language_model.model.layers.23.cross_attn.o_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 406 |
"language_model.model.layers.23.cross_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 407 |
-
"language_model.model.layers.23.cross_attn.o_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 408 |
"language_model.model.layers.23.cross_attn.q_norm.weight": "model-00002-of-00003.safetensors",
|
| 409 |
-
"language_model.model.layers.23.cross_attn.q_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 410 |
"language_model.model.layers.23.cross_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 411 |
-
"language_model.model.layers.23.cross_attn.q_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 412 |
-
"language_model.model.layers.23.cross_attn.v_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 413 |
"language_model.model.layers.23.cross_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 414 |
-
"language_model.model.layers.23.cross_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 415 |
"language_model.model.layers.23.cross_attn_attn_gate": "model-00002-of-00003.safetensors",
|
| 416 |
"language_model.model.layers.23.cross_attn_mlp_gate": "model-00002-of-00003.safetensors",
|
| 417 |
"language_model.model.layers.23.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
|
@@ -522,20 +495,11 @@
|
|
| 522 |
"language_model.model.layers.27.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 523 |
"language_model.model.layers.27.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 524 |
"language_model.model.layers.28.cross_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
| 525 |
-
"language_model.model.layers.28.cross_attn.k_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 526 |
"language_model.model.layers.28.cross_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 527 |
-
"language_model.model.layers.28.cross_attn.k_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 528 |
-
"language_model.model.layers.28.cross_attn.kv_scale": "model-00002-of-00003.safetensors",
|
| 529 |
-
"language_model.model.layers.28.cross_attn.o_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 530 |
"language_model.model.layers.28.cross_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 531 |
-
"language_model.model.layers.28.cross_attn.o_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 532 |
"language_model.model.layers.28.cross_attn.q_norm.weight": "model-00002-of-00003.safetensors",
|
| 533 |
-
"language_model.model.layers.28.cross_attn.q_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 534 |
"language_model.model.layers.28.cross_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 535 |
-
"language_model.model.layers.28.cross_attn.q_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 536 |
-
"language_model.model.layers.28.cross_attn.v_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 537 |
"language_model.model.layers.28.cross_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 538 |
-
"language_model.model.layers.28.cross_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 539 |
"language_model.model.layers.28.cross_attn_attn_gate": "model-00002-of-00003.safetensors",
|
| 540 |
"language_model.model.layers.28.cross_attn_mlp_gate": "model-00002-of-00003.safetensors",
|
| 541 |
"language_model.model.layers.28.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
|
@@ -574,20 +538,11 @@
|
|
| 574 |
"language_model.model.layers.29.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 575 |
"language_model.model.layers.29.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 576 |
"language_model.model.layers.3.cross_attn.k_norm.weight": "model-00001-of-00003.safetensors",
|
| 577 |
-
"language_model.model.layers.3.cross_attn.k_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 578 |
"language_model.model.layers.3.cross_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 579 |
-
"language_model.model.layers.3.cross_attn.k_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 580 |
-
"language_model.model.layers.3.cross_attn.kv_scale": "model-00001-of-00003.safetensors",
|
| 581 |
-
"language_model.model.layers.3.cross_attn.o_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 582 |
"language_model.model.layers.3.cross_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 583 |
-
"language_model.model.layers.3.cross_attn.o_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 584 |
"language_model.model.layers.3.cross_attn.q_norm.weight": "model-00001-of-00003.safetensors",
|
| 585 |
-
"language_model.model.layers.3.cross_attn.q_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 586 |
"language_model.model.layers.3.cross_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 587 |
-
"language_model.model.layers.3.cross_attn.q_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 588 |
-
"language_model.model.layers.3.cross_attn.v_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 589 |
"language_model.model.layers.3.cross_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 590 |
-
"language_model.model.layers.3.cross_attn.v_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 591 |
"language_model.model.layers.3.cross_attn_attn_gate": "model-00001-of-00003.safetensors",
|
| 592 |
"language_model.model.layers.3.cross_attn_mlp_gate": "model-00001-of-00003.safetensors",
|
| 593 |
"language_model.model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
|
@@ -625,17 +580,17 @@
|
|
| 625 |
"language_model.model.layers.30.self_attn.v_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 626 |
"language_model.model.layers.30.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 627 |
"language_model.model.layers.30.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 628 |
-
"language_model.model.layers.31.input_layernorm.weight": "model-
|
| 629 |
-
"language_model.model.layers.31.mlp.down_proj.input_scale": "model-
|
| 630 |
-
"language_model.model.layers.31.mlp.down_proj.weight": "model-
|
| 631 |
-
"language_model.model.layers.31.mlp.down_proj.weight_scale": "model-
|
| 632 |
"language_model.model.layers.31.mlp.gate_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 633 |
"language_model.model.layers.31.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 634 |
"language_model.model.layers.31.mlp.gate_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 635 |
"language_model.model.layers.31.mlp.up_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 636 |
"language_model.model.layers.31.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 637 |
"language_model.model.layers.31.mlp.up_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 638 |
-
"language_model.model.layers.31.post_attention_layernorm.weight": "model-
|
| 639 |
"language_model.model.layers.31.self_attn.k_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 640 |
"language_model.model.layers.31.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 641 |
"language_model.model.layers.31.self_attn.k_proj.weight_scale": "model-00002-of-00003.safetensors",
|
|
@@ -649,47 +604,38 @@
|
|
| 649 |
"language_model.model.layers.31.self_attn.v_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 650 |
"language_model.model.layers.31.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 651 |
"language_model.model.layers.31.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 652 |
-
"language_model.model.layers.32.input_layernorm.weight": "model-
|
| 653 |
-
"language_model.model.layers.32.mlp.down_proj.input_scale": "model-
|
| 654 |
-
"language_model.model.layers.32.mlp.down_proj.weight": "model-
|
| 655 |
-
"language_model.model.layers.32.mlp.down_proj.weight_scale": "model-
|
| 656 |
-
"language_model.model.layers.32.mlp.gate_proj.input_scale": "model-
|
| 657 |
-
"language_model.model.layers.32.mlp.gate_proj.weight": "model-
|
| 658 |
-
"language_model.model.layers.32.mlp.gate_proj.weight_scale": "model-
|
| 659 |
-
"language_model.model.layers.32.mlp.up_proj.input_scale": "model-
|
| 660 |
-
"language_model.model.layers.32.mlp.up_proj.weight": "model-
|
| 661 |
-
"language_model.model.layers.32.mlp.up_proj.weight_scale": "model-
|
| 662 |
-
"language_model.model.layers.32.post_attention_layernorm.weight": "model-
|
| 663 |
-
"language_model.model.layers.32.self_attn.k_proj.input_scale": "model-
|
| 664 |
-
"language_model.model.layers.32.self_attn.k_proj.weight": "model-
|
| 665 |
-
"language_model.model.layers.32.self_attn.k_proj.weight_scale": "model-
|
| 666 |
-
"language_model.model.layers.32.self_attn.kv_scale": "model-
|
| 667 |
-
"language_model.model.layers.32.self_attn.o_proj.input_scale": "model-
|
| 668 |
-
"language_model.model.layers.32.self_attn.o_proj.weight": "model-
|
| 669 |
-
"language_model.model.layers.32.self_attn.o_proj.weight_scale": "model-
|
| 670 |
-
"language_model.model.layers.32.self_attn.q_proj.input_scale": "model-
|
| 671 |
-
"language_model.model.layers.32.self_attn.q_proj.weight": "model-
|
| 672 |
-
"language_model.model.layers.32.self_attn.q_proj.weight_scale": "model-
|
| 673 |
-
"language_model.model.layers.32.self_attn.v_proj.input_scale": "model-
|
| 674 |
-
"language_model.model.layers.32.self_attn.v_proj.weight": "model-
|
| 675 |
-
"language_model.model.layers.32.self_attn.v_proj.weight_scale": "model-
|
| 676 |
"language_model.model.layers.33.cross_attn.k_norm.weight": "model-00003-of-00003.safetensors",
|
| 677 |
-
"language_model.model.layers.33.cross_attn.k_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 678 |
"language_model.model.layers.33.cross_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 679 |
-
"language_model.model.layers.33.cross_attn.k_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 680 |
-
"language_model.model.layers.33.cross_attn.kv_scale": "model-00002-of-00003.safetensors",
|
| 681 |
-
"language_model.model.layers.33.cross_attn.o_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 682 |
"language_model.model.layers.33.cross_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 683 |
-
"language_model.model.layers.33.cross_attn.o_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 684 |
"language_model.model.layers.33.cross_attn.q_norm.weight": "model-00003-of-00003.safetensors",
|
| 685 |
-
"language_model.model.layers.33.cross_attn.q_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 686 |
"language_model.model.layers.33.cross_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 687 |
-
"language_model.model.layers.33.cross_attn.q_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 688 |
-
"language_model.model.layers.33.cross_attn.v_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 689 |
"language_model.model.layers.33.cross_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 690 |
-
"language_model.model.layers.33.
|
| 691 |
-
"language_model.model.layers.33.
|
| 692 |
-
"language_model.model.layers.33.cross_attn_mlp_gate": "model-00002-of-00003.safetensors",
|
| 693 |
"language_model.model.layers.33.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 694 |
"language_model.model.layers.33.mlp.down_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 695 |
"language_model.model.layers.33.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
|
@@ -798,20 +744,11 @@
|
|
| 798 |
"language_model.model.layers.37.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 799 |
"language_model.model.layers.37.self_attn.v_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 800 |
"language_model.model.layers.38.cross_attn.k_norm.weight": "model-00003-of-00003.safetensors",
|
| 801 |
-
"language_model.model.layers.38.cross_attn.k_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 802 |
"language_model.model.layers.38.cross_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 803 |
-
"language_model.model.layers.38.cross_attn.k_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 804 |
-
"language_model.model.layers.38.cross_attn.kv_scale": "model-00003-of-00003.safetensors",
|
| 805 |
-
"language_model.model.layers.38.cross_attn.o_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 806 |
"language_model.model.layers.38.cross_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 807 |
-
"language_model.model.layers.38.cross_attn.o_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 808 |
"language_model.model.layers.38.cross_attn.q_norm.weight": "model-00003-of-00003.safetensors",
|
| 809 |
-
"language_model.model.layers.38.cross_attn.q_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 810 |
"language_model.model.layers.38.cross_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 811 |
-
"language_model.model.layers.38.cross_attn.q_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 812 |
-
"language_model.model.layers.38.cross_attn.v_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 813 |
"language_model.model.layers.38.cross_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 814 |
-
"language_model.model.layers.38.cross_attn.v_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 815 |
"language_model.model.layers.38.cross_attn_attn_gate": "model-00003-of-00003.safetensors",
|
| 816 |
"language_model.model.layers.38.cross_attn_mlp_gate": "model-00003-of-00003.safetensors",
|
| 817 |
"language_model.model.layers.38.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
|
@@ -946,20 +883,11 @@
|
|
| 946 |
"language_model.model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 947 |
"language_model.model.layers.7.self_attn.v_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 948 |
"language_model.model.layers.8.cross_attn.k_norm.weight": "model-00001-of-00003.safetensors",
|
| 949 |
-
"language_model.model.layers.8.cross_attn.k_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 950 |
"language_model.model.layers.8.cross_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 951 |
-
"language_model.model.layers.8.cross_attn.k_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 952 |
-
"language_model.model.layers.8.cross_attn.kv_scale": "model-00001-of-00003.safetensors",
|
| 953 |
-
"language_model.model.layers.8.cross_attn.o_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 954 |
"language_model.model.layers.8.cross_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
| 955 |
-
"language_model.model.layers.8.cross_attn.o_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 956 |
"language_model.model.layers.8.cross_attn.q_norm.weight": "model-00001-of-00003.safetensors",
|
| 957 |
-
"language_model.model.layers.8.cross_attn.q_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 958 |
"language_model.model.layers.8.cross_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
| 959 |
-
"language_model.model.layers.8.cross_attn.q_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 960 |
-
"language_model.model.layers.8.cross_attn.v_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 961 |
"language_model.model.layers.8.cross_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 962 |
-
"language_model.model.layers.8.cross_attn.v_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 963 |
"language_model.model.layers.8.cross_attn_attn_gate": "model-00001-of-00003.safetensors",
|
| 964 |
"language_model.model.layers.8.cross_attn_mlp_gate": "model-00001-of-00003.safetensors",
|
| 965 |
"language_model.model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
|
@@ -973,17 +901,17 @@
|
|
| 973 |
"language_model.model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 974 |
"language_model.model.layers.8.mlp.up_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 975 |
"language_model.model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 976 |
-
"language_model.model.layers.9.input_layernorm.weight": "model-
|
| 977 |
-
"language_model.model.layers.9.mlp.down_proj.input_scale": "model-
|
| 978 |
-
"language_model.model.layers.9.mlp.down_proj.weight": "model-
|
| 979 |
-
"language_model.model.layers.9.mlp.down_proj.weight_scale": "model-
|
| 980 |
"language_model.model.layers.9.mlp.gate_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 981 |
"language_model.model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 982 |
"language_model.model.layers.9.mlp.gate_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 983 |
"language_model.model.layers.9.mlp.up_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 984 |
"language_model.model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 985 |
"language_model.model.layers.9.mlp.up_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 986 |
-
"language_model.model.layers.9.post_attention_layernorm.weight": "model-
|
| 987 |
"language_model.model.layers.9.self_attn.k_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 988 |
"language_model.model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 989 |
"language_model.model.layers.9.self_attn.k_proj.weight_scale": "model-00001-of-00003.safetensors",
|
|
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
+
"total_size": 12951835782
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"language_model.lm_head.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 64 |
"language_model.model.layers.10.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 65 |
"language_model.model.layers.10.mlp.up_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 66 |
"language_model.model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 67 |
+
"language_model.model.layers.10.self_attn.k_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 68 |
+
"language_model.model.layers.10.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 69 |
+
"language_model.model.layers.10.self_attn.k_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 70 |
+
"language_model.model.layers.10.self_attn.kv_scale": "model-00002-of-00003.safetensors",
|
| 71 |
"language_model.model.layers.10.self_attn.o_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 72 |
"language_model.model.layers.10.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
| 73 |
"language_model.model.layers.10.self_attn.o_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 74 |
+
"language_model.model.layers.10.self_attn.q_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 75 |
+
"language_model.model.layers.10.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
| 76 |
+
"language_model.model.layers.10.self_attn.q_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 77 |
+
"language_model.model.layers.10.self_attn.v_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 78 |
+
"language_model.model.layers.10.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 79 |
+
"language_model.model.layers.10.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 80 |
"language_model.model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 81 |
"language_model.model.layers.11.mlp.down_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 82 |
"language_model.model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 126 |
"language_model.model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 127 |
"language_model.model.layers.12.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 128 |
"language_model.model.layers.13.cross_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 129 |
"language_model.model.layers.13.cross_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
|
|
|
|
|
|
| 130 |
"language_model.model.layers.13.cross_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 131 |
"language_model.model.layers.13.cross_attn.q_norm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 132 |
"language_model.model.layers.13.cross_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
|
|
|
| 133 |
"language_model.model.layers.13.cross_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 134 |
"language_model.model.layers.13.cross_attn_attn_gate": "model-00002-of-00003.safetensors",
|
| 135 |
"language_model.model.layers.13.cross_attn_mlp_gate": "model-00002-of-00003.safetensors",
|
| 136 |
"language_model.model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 241 |
"language_model.model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 242 |
"language_model.model.layers.17.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 243 |
"language_model.model.layers.18.cross_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 244 |
"language_model.model.layers.18.cross_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
|
|
|
|
|
|
| 245 |
"language_model.model.layers.18.cross_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 246 |
"language_model.model.layers.18.cross_attn.q_norm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 247 |
"language_model.model.layers.18.cross_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
|
|
|
| 248 |
"language_model.model.layers.18.cross_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 249 |
"language_model.model.layers.18.cross_attn_attn_gate": "model-00002-of-00003.safetensors",
|
| 250 |
"language_model.model.layers.18.cross_attn_mlp_gate": "model-00002-of-00003.safetensors",
|
| 251 |
"language_model.model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 380 |
"language_model.model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 381 |
"language_model.model.layers.22.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 382 |
"language_model.model.layers.23.cross_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 383 |
"language_model.model.layers.23.cross_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
|
|
|
|
|
|
| 384 |
"language_model.model.layers.23.cross_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 385 |
"language_model.model.layers.23.cross_attn.q_norm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 386 |
"language_model.model.layers.23.cross_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
|
|
|
| 387 |
"language_model.model.layers.23.cross_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 388 |
"language_model.model.layers.23.cross_attn_attn_gate": "model-00002-of-00003.safetensors",
|
| 389 |
"language_model.model.layers.23.cross_attn_mlp_gate": "model-00002-of-00003.safetensors",
|
| 390 |
"language_model.model.layers.23.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 495 |
"language_model.model.layers.27.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 496 |
"language_model.model.layers.27.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 497 |
"language_model.model.layers.28.cross_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 498 |
"language_model.model.layers.28.cross_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
|
|
|
|
|
|
| 499 |
"language_model.model.layers.28.cross_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 500 |
"language_model.model.layers.28.cross_attn.q_norm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 501 |
"language_model.model.layers.28.cross_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
|
|
|
| 502 |
"language_model.model.layers.28.cross_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 503 |
"language_model.model.layers.28.cross_attn_attn_gate": "model-00002-of-00003.safetensors",
|
| 504 |
"language_model.model.layers.28.cross_attn_mlp_gate": "model-00002-of-00003.safetensors",
|
| 505 |
"language_model.model.layers.28.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
|
|
|
| 538 |
"language_model.model.layers.29.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 539 |
"language_model.model.layers.29.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 540 |
"language_model.model.layers.3.cross_attn.k_norm.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 541 |
"language_model.model.layers.3.cross_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
|
|
|
|
|
|
|
|
|
| 542 |
"language_model.model.layers.3.cross_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 543 |
"language_model.model.layers.3.cross_attn.q_norm.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 544 |
"language_model.model.layers.3.cross_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
|
|
|
|
|
|
| 545 |
"language_model.model.layers.3.cross_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 546 |
"language_model.model.layers.3.cross_attn_attn_gate": "model-00001-of-00003.safetensors",
|
| 547 |
"language_model.model.layers.3.cross_attn_mlp_gate": "model-00001-of-00003.safetensors",
|
| 548 |
"language_model.model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 580 |
"language_model.model.layers.30.self_attn.v_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 581 |
"language_model.model.layers.30.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 582 |
"language_model.model.layers.30.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 583 |
+
"language_model.model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 584 |
+
"language_model.model.layers.31.mlp.down_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 585 |
+
"language_model.model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 586 |
+
"language_model.model.layers.31.mlp.down_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 587 |
"language_model.model.layers.31.mlp.gate_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 588 |
"language_model.model.layers.31.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
| 589 |
"language_model.model.layers.31.mlp.gate_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 590 |
"language_model.model.layers.31.mlp.up_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 591 |
"language_model.model.layers.31.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
| 592 |
"language_model.model.layers.31.mlp.up_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 593 |
+
"language_model.model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 594 |
"language_model.model.layers.31.self_attn.k_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 595 |
"language_model.model.layers.31.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
| 596 |
"language_model.model.layers.31.self_attn.k_proj.weight_scale": "model-00002-of-00003.safetensors",
|
|
|
|
| 604 |
"language_model.model.layers.31.self_attn.v_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 605 |
"language_model.model.layers.31.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
| 606 |
"language_model.model.layers.31.self_attn.v_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 607 |
+
"language_model.model.layers.32.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 608 |
+
"language_model.model.layers.32.mlp.down_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 609 |
+
"language_model.model.layers.32.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
| 610 |
+
"language_model.model.layers.32.mlp.down_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 611 |
+
"language_model.model.layers.32.mlp.gate_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 612 |
+
"language_model.model.layers.32.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
| 613 |
+
"language_model.model.layers.32.mlp.gate_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 614 |
+
"language_model.model.layers.32.mlp.up_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 615 |
+
"language_model.model.layers.32.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
| 616 |
+
"language_model.model.layers.32.mlp.up_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 617 |
+
"language_model.model.layers.32.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 618 |
+
"language_model.model.layers.32.self_attn.k_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 619 |
+
"language_model.model.layers.32.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
| 620 |
+
"language_model.model.layers.32.self_attn.k_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 621 |
+
"language_model.model.layers.32.self_attn.kv_scale": "model-00003-of-00003.safetensors",
|
| 622 |
+
"language_model.model.layers.32.self_attn.o_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 623 |
+
"language_model.model.layers.32.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
| 624 |
+
"language_model.model.layers.32.self_attn.o_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 625 |
+
"language_model.model.layers.32.self_attn.q_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 626 |
+
"language_model.model.layers.32.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
| 627 |
+
"language_model.model.layers.32.self_attn.q_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 628 |
+
"language_model.model.layers.32.self_attn.v_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 629 |
+
"language_model.model.layers.32.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 630 |
+
"language_model.model.layers.32.self_attn.v_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 631 |
"language_model.model.layers.33.cross_attn.k_norm.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 632 |
"language_model.model.layers.33.cross_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
|
|
|
|
|
|
|
|
|
| 633 |
"language_model.model.layers.33.cross_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 634 |
"language_model.model.layers.33.cross_attn.q_norm.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 635 |
"language_model.model.layers.33.cross_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
|
|
|
|
|
|
| 636 |
"language_model.model.layers.33.cross_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 637 |
+
"language_model.model.layers.33.cross_attn_attn_gate": "model-00003-of-00003.safetensors",
|
| 638 |
+
"language_model.model.layers.33.cross_attn_mlp_gate": "model-00003-of-00003.safetensors",
|
|
|
|
| 639 |
"language_model.model.layers.33.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 640 |
"language_model.model.layers.33.mlp.down_proj.input_scale": "model-00003-of-00003.safetensors",
|
| 641 |
"language_model.model.layers.33.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 744 |
"language_model.model.layers.37.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
| 745 |
"language_model.model.layers.37.self_attn.v_proj.weight_scale": "model-00003-of-00003.safetensors",
|
| 746 |
"language_model.model.layers.38.cross_attn.k_norm.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 747 |
"language_model.model.layers.38.cross_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
|
|
|
|
|
|
|
|
|
| 748 |
"language_model.model.layers.38.cross_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 749 |
"language_model.model.layers.38.cross_attn.q_norm.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 750 |
"language_model.model.layers.38.cross_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
|
|
|
|
|
|
| 751 |
"language_model.model.layers.38.cross_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 752 |
"language_model.model.layers.38.cross_attn_attn_gate": "model-00003-of-00003.safetensors",
|
| 753 |
"language_model.model.layers.38.cross_attn_mlp_gate": "model-00003-of-00003.safetensors",
|
| 754 |
"language_model.model.layers.38.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
|
|
|
| 883 |
"language_model.model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
| 884 |
"language_model.model.layers.7.self_attn.v_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 885 |
"language_model.model.layers.8.cross_attn.k_norm.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 886 |
"language_model.model.layers.8.cross_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
|
|
|
|
|
|
|
|
|
| 887 |
"language_model.model.layers.8.cross_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 888 |
"language_model.model.layers.8.cross_attn.q_norm.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 889 |
"language_model.model.layers.8.cross_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
|
|
|
|
|
|
| 890 |
"language_model.model.layers.8.cross_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 891 |
"language_model.model.layers.8.cross_attn_attn_gate": "model-00001-of-00003.safetensors",
|
| 892 |
"language_model.model.layers.8.cross_attn_mlp_gate": "model-00001-of-00003.safetensors",
|
| 893 |
"language_model.model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
|
|
|
| 901 |
"language_model.model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 902 |
"language_model.model.layers.8.mlp.up_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 903 |
"language_model.model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 904 |
+
"language_model.model.layers.9.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 905 |
+
"language_model.model.layers.9.mlp.down_proj.input_scale": "model-00002-of-00003.safetensors",
|
| 906 |
+
"language_model.model.layers.9.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
| 907 |
+
"language_model.model.layers.9.mlp.down_proj.weight_scale": "model-00002-of-00003.safetensors",
|
| 908 |
"language_model.model.layers.9.mlp.gate_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 909 |
"language_model.model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
| 910 |
"language_model.model.layers.9.mlp.gate_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 911 |
"language_model.model.layers.9.mlp.up_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 912 |
"language_model.model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
| 913 |
"language_model.model.layers.9.mlp.up_proj.weight_scale": "model-00001-of-00003.safetensors",
|
| 914 |
+
"language_model.model.layers.9.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 915 |
"language_model.model.layers.9.self_attn.k_proj.input_scale": "model-00001-of-00003.safetensors",
|
| 916 |
"language_model.model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
| 917 |
"language_model.model.layers.9.self_attn.k_proj.weight_scale": "model-00001-of-00003.safetensors",
|
preprocessor_config.json
CHANGED
|
@@ -16,6 +16,7 @@
|
|
| 16 |
0.27577711
|
| 17 |
],
|
| 18 |
"max_image_tiles": 4,
|
|
|
|
| 19 |
"resample": 2,
|
| 20 |
"rescale_factor": 0.00392156862745098,
|
| 21 |
"size": {
|
|
|
|
| 16 |
0.27577711
|
| 17 |
],
|
| 18 |
"max_image_tiles": 4,
|
| 19 |
+
"processor_class": "MllamaProcessor",
|
| 20 |
"resample": 2,
|
| 21 |
"rescale_factor": 0.00392156862745098,
|
| 22 |
"size": {
|