Phase 4: OpenClaw agent setup + expanded model catalog

- Add OpenClaw launchd plist (gateway on port 8080)
- Update Llama-3.3-70B Modelfile: fix FROM path, add tool-calling TEMPLATE,
  set num_ctx 32768 (fits 70B in 64GB with safe headroom)
- Update Codestral-22B and Qwen3-32B Modelfiles
- Add Modelfiles for all models in ~/Models/LLM and ~/Models/MLX:
  EVA-LLaMA-3.33-70B, Midnight-Miqu-70B, QwQ-32B, Qwen3.5-35B,
  Qwen3-Coder-30B, Qwen3-Coder-Next, Qwen3-VL-30B, GLM-4.6V-Flash,
  DeepSeek-R1-8B, gemma-3-27b, and MLX variants
- Add import-local-models.sh helper script

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
This commit is contained in:
Aodhan Collins
2026-03-06 00:03:06 +00:00
parent a911693057
commit c4f3dbed77
85 changed files with 8845846 additions and 3 deletions

View File

@@ -0,0 +1,117 @@
{% macro render_extra_keys(json_dict, handled_keys) %}
{%- if json_dict is mapping %}
{%- for json_key in json_dict if json_key not in handled_keys %}
{%- if json_dict[json_key] is string %}
{{-'\n<' ~ json_key ~ '>' ~ (json_dict[json_key] | string) ~ '</' ~ json_key ~ '>' }}
{%- else %}
{{- '\n<' ~ json_key ~ '>' ~ (json_dict[json_key] | tojson) ~ '</' ~ json_key ~ '>' }}
{%- endif %}
{%- endfor %}
{%- endif %}
{%- endmacro %}
{%- if messages[0]["role"] == "system" %}
{%- set system_message = messages[0]["content"] %}
{%- set loop_messages = messages[1:] %}
{%- else %}
{%- set loop_messages = messages %}
{%- endif %}
{%- if not tools is defined %}
{%- set tools = [] %}
{%- endif %}
{%- if system_message is defined %}
{{- "<|im_start|>system\n" + system_message }}
{%- else %}
{%- if tools is iterable and tools | length > 0 %}
{{- "<|im_start|>system\nYou are Qwen, a helpful AI assistant that can interact with a computer to solve tasks." }}
{%- endif %}
{%- endif %}
{%- if tools is iterable and tools | length > 0 %}
{{- "\n\n# Tools\n\nYou have access to the following functions:\n\n" }}
{{- "<tools>" }}
{%- for tool in tools %}
{%- if tool.function is defined %}
{%- set tool = tool.function %}
{%- endif %}
{{- "\n<function>\n<name>" ~ tool.name ~ "</name>" }}
{%- if tool.description is defined %}
{{- '\n<description>' ~ (tool.description | trim) ~ '</description>' }}
{%- endif %}
{{- '\n<parameters>' }}
{%- if tool.parameters is defined and tool.parameters is mapping and tool.parameters.properties is defined and tool.parameters.properties is mapping %}
{%- for param_name, param_fields in tool.parameters.properties|items %}
{{- '\n<parameter>' }}
{{- '\n<name>' ~ param_name ~ '</name>' }}
{%- if param_fields.type is defined %}
{{- '\n<type>' ~ (param_fields.type | string) ~ '</type>' }}
{%- endif %}
{%- if param_fields.description is defined %}
{{- '\n<description>' ~ (param_fields.description | trim) ~ '</description>' }}
{%- endif %}
{%- set handled_keys = ['name', 'type', 'description'] %}
{{- render_extra_keys(param_fields, handled_keys) }}
{{- '\n</parameter>' }}
{%- endfor %}
{%- endif %}
{%- set handled_keys = ['type', 'properties'] %}
{{- render_extra_keys(tool.parameters, handled_keys) }}
{{- '\n</parameters>' }}
{%- set handled_keys = ['type', 'name', 'description', 'parameters'] %}
{{- render_extra_keys(tool, handled_keys) }}
{{- '\n</function>' }}
{%- endfor %}
{{- "\n</tools>" }}
{{- '\n\nIf you choose to call a function ONLY reply in the following format with NO suffix:\n\n<tool_call>\n<function=example_function_name>\n<parameter=example_parameter_1>\nvalue_1\n</parameter>\n<parameter=example_parameter_2>\nThis is the value for the second parameter\nthat can span\nmultiple lines\n</parameter>\n</function>\n</tool_call>\n\n<IMPORTANT>\nReminder:\n- Function calls MUST follow the specified format: an inner <function=...></function> block must be nested within <tool_call></tool_call> XML tags\n- Required parameters MUST be specified\n- You may provide optional reasoning for your function call in natural language BEFORE the function call, but NOT after\n- If there is no function call available, answer the question like normal with your current knowledge and do not tell the user about function calls\n</IMPORTANT>' }}
{%- endif %}
{%- if system_message is defined %}
{{- '<|im_end|>\n' }}
{%- else %}
{%- if tools is iterable and tools | length > 0 %}
{{- '<|im_end|>\n' }}
{%- endif %}
{%- endif %}
{%- for message in loop_messages %}
{%- if message.role == "assistant" and message.tool_calls is defined and message.tool_calls is iterable and message.tool_calls | length > 0 %}
{{- '<|im_start|>' + message.role }}
{%- if message.content is defined and message.content is string and message.content | trim | length > 0 %}
{{- '\n' + message.content | trim + '\n' }}
{%- endif %}
{%- for tool_call in message.tool_calls %}
{%- if tool_call.function is defined %}
{%- set tool_call = tool_call.function %}
{%- endif %}
{{- '\n<tool_call>\n<function=' + tool_call.name + '>\n' }}
{%- if tool_call.arguments is defined %}
{%- for args_name, args_value in tool_call.arguments|items %}
{{- '<parameter=' + args_name + '>\n' }}
{%- set args_value = args_value if args_value is string else args_value | tojson %}
{{- args_value }}
{{- '\n</parameter>\n' }}
{%- endfor %}
{%- endif %}
{{- '</function>\n</tool_call>' }}
{%- endfor %}
{{- '<|im_end|>\n' }}
{%- elif message.role == "user" or message.role == "system" or message.role == "assistant" %}
{{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
{%- elif message.role == "tool" %}
{%- if loop.previtem and loop.previtem.role != "tool" %}
{{- '<|im_start|>user' }}
{%- endif %}
{{- '\n<tool_response>\n' }}
{{- message.content }}
{{- '\n</tool_response>' }}
{%- if not loop.last and loop.nextitem.role != "tool" %}
{{- '<|im_end|>\n' }}
{%- elif loop.last %}
{{- '<|im_end|>\n' }}
{%- endif %}
{%- else %}
{{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>\n' }}
{%- endif %}
{%- endfor %}
{%- if add_generation_prompt %}
{{- '<|im_start|>assistant\n' }}
{%- endif %}

View File

@@ -0,0 +1,825 @@
{
"architectures": [
"Qwen3NextForCausalLM"
],
"attention_bias": false,
"attention_dropout": 0,
"bos_token_id": 151643,
"decoder_sparse_step": 1,
"eos_token_id": [
151645,
151643
],
"full_attention_interval": 4,
"head_dim": 256,
"hidden_act": "silu",
"hidden_size": 2048,
"initializer_range": 0.02,
"intermediate_size": 5120,
"linear_conv_kernel_dim": 4,
"linear_key_head_dim": 128,
"linear_num_key_heads": 16,
"linear_num_value_heads": 32,
"linear_value_head_dim": 128,
"max_position_embeddings": 262144,
"mlp_only_layers": [],
"model_type": "qwen3_next",
"moe_intermediate_size": 512,
"norm_topk_prob": true,
"num_attention_heads": 16,
"num_experts": 512,
"num_experts_per_tok": 10,
"num_hidden_layers": 48,
"num_key_value_heads": 2,
"output_router_logits": false,
"partial_rotary_factor": 0.25,
"quantization": {
"group_size": 64,
"bits": 4,
"mode": "affine",
"model.layers.0.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.0.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.1.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.1.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.2.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.2.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.3.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.3.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.4.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.4.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.5.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.5.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.6.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.6.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.7.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.7.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.8.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.8.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.9.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.9.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.10.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.10.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.11.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.11.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.12.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.12.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.13.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.13.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.14.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.14.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.15.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.15.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.16.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.16.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.17.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.17.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.18.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.18.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.19.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.19.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.20.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.20.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.21.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.21.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.22.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.22.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.23.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.23.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.24.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.24.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.25.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.25.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.26.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.26.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.27.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.27.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.28.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.28.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.29.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.29.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.30.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.30.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.31.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.31.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.32.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.32.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.33.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.33.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.34.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.34.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.35.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.35.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.36.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.36.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.37.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.37.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.38.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.38.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.39.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.39.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.40.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.40.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.41.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.41.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.42.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.42.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.43.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.43.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.44.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.44.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.45.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.45.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.46.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.46.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.47.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.47.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
}
},
"quantization_config": {
"group_size": 64,
"bits": 4,
"mode": "affine",
"model.layers.0.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.0.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.1.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.1.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.2.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.2.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.3.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.3.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.4.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.4.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.5.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.5.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.6.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.6.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.7.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.7.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.8.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.8.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.9.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.9.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.10.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.10.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.11.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.11.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.12.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.12.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.13.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.13.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.14.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.14.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.15.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.15.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.16.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.16.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.17.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.17.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.18.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.18.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.19.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.19.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.20.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.20.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.21.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.21.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.22.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.22.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.23.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.23.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.24.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.24.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.25.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.25.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.26.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.26.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.27.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.27.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.28.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.28.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.29.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.29.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.30.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.30.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.31.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.31.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.32.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.32.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.33.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.33.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.34.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.34.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.35.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.35.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.36.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.36.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.37.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.37.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.38.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.38.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.39.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.39.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.40.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.40.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.41.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.41.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.42.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.42.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.43.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.43.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.44.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.44.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.45.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.45.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.46.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.46.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
},
"model.layers.47.mlp.gate": {
"group_size": 64,
"bits": 8
},
"model.layers.47.mlp.shared_expert_gate": {
"group_size": 64,
"bits": 8
}
},
"rms_norm_eps": 1e-06,
"rope_scaling": null,
"rope_theta": 5000000,
"router_aux_loss_coef": 0.001,
"shared_expert_intermediate_size": 512,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.57.0.dev0",
"use_cache": true,
"use_sliding_window": false,
"vocab_size": 151936
}

View File

@@ -0,0 +1,13 @@
{
"bos_token_id": 151643,
"do_sample": true,
"eos_token_id": [
151645,
151643
],
"pad_token_id": 151643,
"temperature": 1.0,
"top_k": 40,
"top_p": 0.95,
"transformers_version": "4.57.3"
}

File diff suppressed because it is too large Load Diff

File diff suppressed because it is too large Load Diff

File diff suppressed because one or more lines are too long