- Blog
- Qwen3 Release: Alibaba's Open-Source AI Model Rivals GPT-4 and Claude
Qwen3 Release: Alibaba's Open-Source AI Model Rivals GPT-4 and Claude
Qwen3 Release: Alibaba's Open-Source AI Model Rivals GPT-4 and Claude
In late April 2025, Alibaba’s Qwen team released Qwen3, a new family of open-source large language models (LLMs). Qwen3 spans eight models from 0.6 billion to 235 billion parameters, including two mixture-of-experts (MoE) variants (a 235B model with 22B “active” parameters and a 30B model with 3B active) and six dense models. Under an Apache 2.0 license, these models are freely available (e.g. on Hugging Face and Alibaba’s Qwen Chat) for research or deployment. Alibaba claims Qwen3 matches or exceeds top AI models on key tasks: for example, the flagship Qwen3-235B-A22B outperforms OpenAI’s o1 model and DeepSeek’s R1 on coding/math benchmarks, and even the small 4B Qwen3 rivals the much larger Qwen2.5-72B. In short, the Qwen3 release delivers an open, high-performance LLM lineup with new features for hybrid reasoning and broad language support.
🔍 Key Features and Architecture
Hybrid Thinking Modes
Qwen3 introduces a toggleable reasoning mode. In “Thinking Mode” the model takes extra steps (step-by-step chain-of-thought) to solve hard problems, while “Non-Thinking Mode” yields quick answers for simple queries. Users can switch modes via prompts or a UI button. This dynamic “thinking budget” lets users balance accuracy vs. speed. (Anthropic’s new Claude 3.7 also offers a similar hybrid mode and even controls the token budget for reasoning.)
Mixture-of-Experts + Dense Models
The Qwen3 family includes two MoE models (235B total params with 22B active, and 30B total with 3B active) and six standard (dense) models (0.6B, 1.7B, 4B, 8B, 14B, 32B). MoE models use “experts” – small specialist sub-models – that activate only when needed, improving efficiency (only ~10% of parameters fire). Even at smaller sizes, Qwen3 is architected for strong reasoning. Notably, context lengths are very large (32K tokens for the 4B model and up to 128K tokens for larger models), far beyond most public LLMs.
Massive Multilingual Training Data
Qwen3 was pretrained on ~36 trillion tokens across 119 languages and dialects. This includes web text, PDF documents (text extracted by Qwen’s older vision-language model), plus synthetic data generated by earlier Qwen models for math and coding (textbooks, Q&A, code snippets). The expanded data scale (double the 18T tokens used for Qwen2.5) and diverse content are credited for big gains in STEM and reasoning skills.
Pretraining and Finetuning
The team used a multi-stage training pipeline. In pretraining, they ran three phases: 30T tokens of general web data (4K context), then 5T tokens enriched with STEM and coding, and finally high-quality long-context data to extend to 32K contexts. In post-training, Qwen3 underwent four steps: long chain-of-thought fine-tuning, reinforcement learning for reasoning, integrating quick-answer examples, and broad RL on multi-domain tasks. This process (inspired by recent “DeepSeek” techniques) yields models that can reason deeply yet respond fast when needed.
Open and Accessible
All Qwen3 models are open-sourced (Apache 2.0). They can be used via the Qwen Chat web/mobile apps, or downloaded for local use (compatible with frameworks like Hugging Face Transformers, Ollama, vLLM, llama.cpp, etc.). Developer tools like Qwen-Agent support building AI agents with Qwen3, and cloud hosts (e.g. Fireworks AI, Hyperbolic) also offer Qwen3 instances. The broad size range (0.6B model even runs on phones) and open license make Qwen3 usable from mobile apps to data centers.
📊 Performance Benchmarks
Alibaba reports that Qwen3’s flagship is very competitive with the latest LLMs. In internal tests: Qwen3-235B-A22B beats OpenAI’s o1 and DeepSeek R1 on coding and math benchmarks, and is on par with Google’s Gemini 2.5-Pro. For example, on the LiveCodeBench coding test, Qwen3-235B scored about 70.7%, trailing only OpenAI’s new o4-mini (80%). On a hard math exam (AIME 2024), it scored 85.7% vs. 94% for o4-mini. The 30B model even outperformed DeepSeek V3 and OpenAI’s GPT-4o (Omni) on certain benchmarks.
Independent media note that on Codeforces programming problems, Qwen3-235B “just beats” OpenAI’s o3-mini and Google’s Gemini 2.5 Pro, and also bests o3-mini on a reasoning test (BFCL). Its largest publicly available model (Qwen3-32B) also surpasses OpenAI’s o1 on coding tasks. In short, Qwen3 sets a new bar for open models, rivaling closed giants in many areas. However, it does not clearly exceed the very top-of-line GPT variants (GPT-4 / o4) or Claude 3/4 on all benchmarks, so these older leaders may remain ahead in some general or creative tasks.
🔁 Comparison with Other LLMs
Model | Key Facts | Strengths | Weaknesses |
---|---|---|---|
Qwen3 (Alibaba) | Open-source (Apache 2.0) LLMs, up to 235B params (22B active); hybrid reasoning modes; supports 119 languages. | Excellent at code/math/complex queries; flexible reasoning-vs-speed; very long context (up to 128K tokens); free to use and modify; strong multilingual support. | Newer/less proven than GPT-4; content filters (Chinese policy restrictions apply); smaller model variants may underperform very best models. |
GPT-4 (OpenAI) | Proprietary (closed) model (huge size, rumored ~175B); up to 32K token context; widely used via API. | State-of-the-art performance across tasks; extensive training data; strong reasoning and creative output; robust safety fine-tuning. | Not open-source (API only); expensive; limited direct customization; no built-in mode toggle (must prompt). |
Claude 3/4 (Anthropic) | Proprietary (closed) models; Claude 3.7 “Sonnet” supports 200K token context and hybrid modes. | Emphasizes safety and alignment; very long context (up to 200K tokens); has built-in standard/extended thinking modes; good at following instructions. | Closed source; generally smaller community; performance slightly below GPT-4 on benchmarks. |
LLaMA 2/3/4 (Meta) | Open-source LLMs, up to ~70B in LLaMA 2 and 405B in LLaMA 3.1; supports fine-tuning. | Fully open and well-supported; strong baseline performance; efficient and easy to run locally. | Lower ceiling than Qwen3’s 235B+MoE; shorter context windows (4K–16K); no special reasoning toggle. |
💡 Use Cases
AI Chat and Assistants
Like ChatGPT, Qwen3 can power chatbots and virtual assistants. Its large context and reasoning toggle make it good for complicated customer queries (e.g. technical support) or storytelling tasks where step-by-step thought helps. Because Qwen3 covers 119 languages, it’s useful for multilingual chatbots and translation services. Alibaba itself embeds Qwen3 in products (e.g. Qwen Chat) and it can be deployed in enterprise apps (e.g. via cloud partners Fireworks AI, Hyperbolic).
Coding and STEM
Qwen3 is optimized for code and math. Benchmarks show its largest models outperform many peers on programming tasks. This makes it well-suited as a developer assistant or code generator (like GitHub Copilot). Educational tools can use Qwen3 to tutor students: its “thinking” mode can walk through hard math or logic problems step by step, improving learning.
Data & Document Processing
The huge context window (up to 128K tokens) means Qwen3 can handle very large documents (essays, legal texts, codebases) in one go. Businesses might use it to summarize lengthy reports, analyze contracts, or answer questions about detailed documents without cutting them into pieces.
Agentic Workflows
With reinforced training and a built-in tool-calling framework, Qwen3 supports “AI agents” that can execute multi-step tasks. For example, an AI agent built on Qwen3 could parse an email (facts extraction), query a database, and draft a response, all in one session. Alibaba’s mention of “tool-calling” and the Qwen-Agent toolkit highlights these advanced applications.
Innovation & Research
As a fully open model, Qwen3 encourages experimentation. Over 100,000 derivative models have already been built on the Qwen platform, making it “the world’s largest open-source AI ecosystem.” Researchers and startups can use Qwen3 to prototype new AI tools without licensing fees or restrictions (unlike closed models).
🧠 Conclusion
The Qwen3 release brings a versatile, high-capability LLM to the AI community. Its combination of open access (Apache-licensed code/weights), powerful reasoning modes, and broad language support means chat users and developers worldwide can “think deeper, act faster” with AI. While GPT-4 and Claude remain state-of-art in many areas, Qwen3 closes the gap and democratizes advanced capabilities. As Alibaba’s team puts it, Qwen3 aims to “empower researchers, developers, and organizations… to build innovative solutions.”