Address
33-17, Q Sentral.
2A, Jalan Stesen Sentral 2, Kuala Lumpur Sentral,
50470 Federal Territory of Kuala Lumpur
Contact
+603-2701-3606
[email protected]
Address
33-17, Q Sentral.
2A, Jalan Stesen Sentral 2, Kuala Lumpur Sentral,
50470 Federal Territory of Kuala Lumpur
Contact
+603-2701-3606
[email protected]
In late April 2025, Alibaba unveiled its Qwen3 series of large language models, marking another leap in China’s rapid AI development following the breakout success of DeepSeek. The new release underscores Alibaba’s commitment to outpacing rivals by offering advanced capabilities and driving down deployment costs.
Building on its predecessor, Qwen3 introduces hybrid reasoning—a blend of conventional pattern-recognition and dynamic logic modules—that boosts adaptability and accuracy in coding, math problem solving, and complex decision workflows. This makes Qwen3 especially suited for enterprise applications like automated code review and data-driven analytics.
Qwen3’s mixture-of-experts (MoE) design splits computation across multiple smaller sub-models, activating only a subset per request. One variant, Qwen3-MoE-15B-A2B, uses 1.5 billion active parameters to match the performance of much larger networks—significantly reducing cloud-compute bills. Smaller 600 million–parameter versions are also available for on-device AI.
Unlike many commercial AI offerings, Qwen3’s weights are open-source under an accessible license, inviting developers and researchers to innovate freely. Since the Qwen family’s debut in April 2023, its models have amassed 40 million downloads on platforms like Hugging Face and inspired over 50,000 derivative models—a testament to strong community adoption.
Alibaba Cloud reports that API pricing for its Qwen series has fallen 97% over the past year, with the cost to process one million tokens now as low as CNY 0.50 (0.007 U.S. cents). This exponential drop in inference expenses positions Qwen3 as one of the most affordable high-performance LLMs on the market.
To support these advances, Alibaba has pledged $53 billion in AI and cloud infrastructure over the next three years, aimed at expanding data-center capacity, accelerating chip R&D, and scaling GPU farms across Asia. This bet underscores how crucial AI is to Alibaba’s future commerce, logistics, and enterprise-cloud ambitions.
DeepSeek’s rapid rise—training world-class models in weeks at a fraction of Western costs—ignited China’s AI arms race. Within days of Qwen3’s debut, Baidu released its Ernie 4.5 Turbo and Ernie X1 Turbo for reasoning-heavy tasks, while startups and incumbents alike scramble to close the gap.
Alibaba’s Qwen3 cements its role in the evolving post-DeepSeek era—delivering smarter reasoning, cost-efficient deployment, and broad community access. As global competition intensifies, this combination of performance, affordability, and open-source collaboration may well define the next chapter of AI-driven innovation.
1. What makes Qwen3 different from Qwen 2.5?
Qwen3 adds hybrid reasoning modules and a mixture-of-experts architecture, improving task adaptability and cutting inference costs by activating only needed parameters.
2. How does Qwen3 compare to DeepSeek’s models?
While both achieve top benchmarks in math and coding, Qwen3 offers broader open-source access and significantly lower deployment expenses thanks to MoE efficiency.
3. Who can use Qwen3 and how?
Anyone can access Qwen3 via Alibaba Cloud’s Model Studio—no local hardware required—and tap into community-driven extensions on platforms like Hugging Face.
Sources Bloomberg