Back to Blog

Qwen 3.5: How Alibaba's Open-Source Models Are Beating GPT and Claude

Qwen 3.5: How Alibaba's Open-Source Models Are Beating GPT and Claude

Alibaba's Qwen 3.5 family delivers frontier-level performance in models small enough to run on a laptop. With 201 language support and Apache 2.0 licensing, the open-source AI race just got a new leader.

In a span of just 16 days, Alibaba's Qwen team shipped nine models across three size tiers, completing one of the most aggressive open-source AI rollouts in history. The Qwen 3.5 family, ranging from 0.8 billion to 122 billion parameters, is not just competitive with proprietary models from OpenAI and Anthropic but is outperforming them on key benchmarks while remaining fully open-source under the Apache 2.0 licence.

Small Models, Big Performance

The headline result is striking: Qwen3.5-9B, a model small enough to run on a standard laptop, scored 81.7 on the GPQA Diamond benchmark, surpassing OpenAI's gpt-oss-120B, a model with over ten times its parameter count. This is not a cherry-picked metric. Across coding, mathematical reasoning, and multilingual understanding benchmarks, the Qwen 3.5 Small models consistently punch well above their weight class. The secret lies in their architecture: a Mixture-of-Experts design that activates only a fraction of total parameters for any given token, combined with a 3:1 ratio of linear to full attention that enables a massive 262,000-token context window while keeping inference efficient.

The Medium Models Take on Frontier Labs

The Qwen3.5 Medium series, including the 35B-A3B, 122B-A10B, and 27B variants, target the sweet spot where most enterprise AI deployments live. These models offer performance comparable to Claude Sonnet 4.5 and GPT-5-mini while being deployable on-premises, a critical requirement for industries with strict data privacy and regulatory compliance obligations. All three Medium models support agentic tool calling natively, making them viable foundations for building autonomous AI systems without relying on proprietary APIs.

201 Languages and the Global AI Gap

Perhaps the most underappreciated feature of Qwen 3.5 is its support for 201 languages and dialects, up from 82 in the previous generation. For markets like India, where multilingual AI capability is essential for reaching users across linguistic boundaries, this breadth of language support unlocks use cases that English-centric models simply cannot serve. At QverLabs, multilingual capability is foundational to how we build products, from our agentic AI platforms to localised compliance tools that need to process regulatory documents in multiple languages.

The Team Exodus Problem

The rollout was not without drama. Within 24 hours of shipping the Qwen 3.5 Small models, the project's technical architect and several other key team members departed Alibaba. This raises legitimate questions about the long-term sustainability of the Qwen project and whether the institutional knowledge required to maintain and evolve these models will survive the talent drain. For organisations building on top of Qwen models, this is a risk factor worth monitoring closely.

What This Means for Enterprise AI

The Qwen 3.5 release accelerates a trend that has been building throughout 2025 and into 2026: open-source models are closing the gap with proprietary ones faster than anyone expected. For enterprise teams, this means more viable options for on-premises deployment, lower inference costs, and reduced vendor lock-in. The trade-off is that open-source models require more engineering effort to deploy, fine-tune, and maintain, making them best suited for organisations with in-house ML expertise or partners who can provide it.