Qwen3-Coder is trained on a large mix of code and text and can handle large inputs. “Trained on an extensive dataset of code and general text data, Qwen3-Coder is engineered for robust agentic coding. It natively supports a context window of 256K tokens, extendable up to 1 million tokens, enabling it to process vast codebases in a single session. Its superior performance stems not only from scaling across tokens, context length, and synthetic data during pre-training, but also from innovative post-training techniques such as long-horizon reinforcement learning (agent RL),” the company said in a statement.
The upgrade lets Qwen3-Coder tackle complex, real-world coding problems via multi-step interactions with external tools. “Qwen3-Coder achieves SOTA performance among open-source models on SWE-Bench Verified (a benchmark for evaluating AI models’ ability to solve real-world software issues), even without test-time or inference scaling.”
Notably, agentic AI coding is transforming software development by enabling more autonomous, efficient, and accessible programming workflows. With its open-source availability, strong agentic coding capabilities, and seamless compatibility with popular developer tools and interfaces, Alibaba aims to position Qwen3-Coder as a valuable tool for global software developers.
The model is available on Hugging Face and GitHub, while developers can also access the model on Qwen Chat or via cost-effective APIs through Model Studio, Alibaba’s generative AI development platform.
Alibaba claimed Qwen-based coding models have already surpassed 20 million downloads globally. Earlier this year, it unveiled AI models like Qwen 2.5 Max and Qwen 3, which claimed to outperform DeepSeek’s R1 model and even the likes of OpenAI’s GPT-4o and Meta’s Llama.
The Qwen3-Coder launch comes amid AI major OpenAI’s recent launch of ChatGPT Agent, a new tool designed to handle complex, multi-step tasks using its own virtual computer.