{ "architectures": [ "Qwen3ForCausalLM" ], "attention_bias": false, "attention_dropout": 0.0, "chat_template": "{%- for message in messages %}\n{%- if message.role == \"user\" %}\n{{- '<|text_start|>\n' + message.content + '<|text_end|>\n' }}\n{%- elif message.role == \"assistant\" %}\n{{- '<|corrected_start|>\n' + message.content + '<|corrected_end|>\n' }}\n{%- else %}\n{# Raise an error for unsupported roles, as per requirement to remove system message stuff #}\n{{- raise('Unknown role: ' + message.role) }}\n{%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n{{- '<|corrected_start|>\n' }}\n{%- endif %}", "eos_token_id": 151643, "head_dim": 128, "hidden_act": "silu", "hidden_size": 2560, "initializer_range": 0.02, "intermediate_size": 9728, "max_position_embeddings": 32768, "max_window_layers": 36, "model_type": "qwen3", "num_attention_heads": 32, "num_hidden_layers": 36, "num_key_value_heads": 8, "pad_token_id": 151654, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 1000000, "sliding_window": null, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "transformers_version": "4.52.4", "unsloth_fixed": true, "unsloth_version": "2025.5.9", "use_cache": true, "use_sliding_window": false, "vocab_size": 151673 }