Alibaba's Qwen Team has made headlines once again with the launch of the Qwen3-Coder-480B-A35B-Instruct, a state-of-the-art open-source language model designed to enhance software development. This new model arrives shortly after the release of the Qwen3-235B-A22B-2507, which quickly garnered attention as the leading non-reasoning large language model globally, outpacing even well-funded proprietary models from tech giants like Google and OpenAI. The Qwen3-Coder is tailored for complex coding tasks, enabling developers to create fully functional applications in mere seconds or minutes. It is set to rival proprietary solutions such as Claude Sonnet-4 in coding efficiency. The model has already achieved impressive benchmark scores among open-source alternatives and is available on platforms like Hugging Face and GitHub, as well as through Alibaba's own Qwen API. One of the standout features of Qwen3-Coder is its open-source Apache 2.0 license, allowing businesses to use, modify, and deploy the model at no cost. This accessibility positions it as a strong competitor in the realm of coding models, particularly for enterprises looking for cost-effective solutions. Sebastian Raschka, a prominent LLM researcher, has praised the model on social media, stating, "This might be the best coding model yet. General-purpose is cool, but if you want the best at coding, specialization wins. No free lunch." Such endorsements from industry experts underscore the model's potential impact. Developers can easily access the model's code through Hugging Face, while those lacking the infrastructure to self-host can utilize the Qwen API, with pricing starting as low as $1 per million tokens for various input/output configurations. The Qwen3-Coder boasts a Mixture-of-Experts architecture featuring 480 billion parameters, with 35 billion active per query, and is capable of managing context lengths of up to 256,000 tokens natively. This allows it to process extensive documents or repositories seamlessly. Its design includes 62 layers and 96 attention heads, optimized for efficiency in instruction-following tasks. Early impressions from AI engineers and developers have been overwhelmingly positive, with many highlighting its superior performance in real-world applications. Kevin Nelson, an educator and AI enthusiast, remarked on social media that "Qwen3-Coder is on another level," citing its ability to execute tasks with remarkable contextual awareness. As the Qwen team works on additional model sizes to cater to a broader audience, they are also exploring the potential for self-improvement, aiming to refine the model's performance through ongoing real-world use. This innovative approach could redefine the landscape of coding models, making Qwen3-Coder a pivotal tool for developers across various industries.
During a recent all-employee town hall, Microsoft leaders provided insights into their new return-to-office (RTO) policy...
Business Insider | Sep 11, 2025, 18:35In a groundbreaking move, Albania has appointed an artificial intelligence bot named Diella as a minister responsible fo...
Mint | Sep 11, 2025, 15:06In a remarkable validation of Stephen Hawking's theories, new research has provided compelling evidence supporting the p...
Ars Technica | Sep 11, 2025, 18:45For seven years, fans have passionately awaited the release of Hollow Knight: Silksong, a sequel that has now arrived ju...
Ars Technica | Sep 11, 2025, 17:50Senator Ted Cruz's latest initiative aimed at reshaping AI regulations has sparked significant backlash from advocates c...
Ars Technica | Sep 11, 2025, 18:25