--- license: apache-2.0 base_model: unsloth/Qwen3-30B-A3B-Instruct-2507 tags: - qwen3 - chat - instruct --- # OpenPipe/Qwen3-30B-A3B-Instruct-2507 This is a copy of [unsloth/Qwen3-30B-A3B-Instruct-2507](https://huggingface.co/unsloth/Qwen3-30B-A3B-Instruct-2507) with a **fixed chat template** for SFT training. ## Changes from the original The only change is the chat template. The original Qwen3 chat template adds `` tags **inconsistently** - only to the last assistant message in a conversation, not to earlier assistant messages in the history. This causes issues during SFT training because: - Historical assistant messages: `<|im_start|>assistant\nHi there!<|im_end|>` - Target assistant message: `<|im_start|>assistant\n\n\n\n\nI am well!<|im_end|>` The model learns an inconsistent pattern, leading to degraded outputs. ## Fixed template This model uses a **consistent** chat template that adds `\n\n\n\n` to ALL assistant messages: ``` {{- '<|im_start|>assistant\n\n\n\n\n' + content }} ``` This ensures the model learns a single, consistent pattern during training. ## Usage Use this model as your base model for SFT training with Qwen3-30B-A3B when you need consistent chat formatting. ```python from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("OpenPipe/Qwen3-30B-A3B-Instruct-2507") model = AutoModelForCausalLM.from_pretrained("OpenPipe/Qwen3-30B-A3B-Instruct-2507") ``` ## Model Details - **Base Model**: [unsloth/Qwen3-30B-A3B-Instruct-2507](https://huggingface.co/unsloth/Qwen3-30B-A3B-Instruct-2507) - **Parameters**: 30B (3B active with MoE) - **License**: Apache 2.0