Alibaba launches more efficient Qwen3-Next AI model
Jinse Finance reported that Alibaba's Tongyi Qianwen has released the next-generation foundational model architecture, Qwen3-Next, and open-sourced the Qwen3-Next-80B-A3B series models based on this architecture. Compared to the MoE model structure of Qwen3, this architecture features the following core improvements: hybrid attention mechanism, high-sparsity MoE structure, a series of training-stable optimizations, and a multi-token prediction mechanism that enhances inference efficiency. Based on the Qwen3-Next model structure, Alibaba has trained the Qwen3-Next-80B-A3B-Base model, which has 80 billion parameters but only activates 3 billion parameters. This Base model achieves performance comparable to or slightly better than the Qwen3-32B dense model, while its training cost (GPU hours) is less than one-tenth of Qwen3-32B, and its inference throughput for contexts above 32k is more than ten times that of Qwen3-32B, achieving exceptional cost-effectiveness in both training and inference.
Disclaimer: The content of this article solely reflects the author's opinion and does not represent the platform in any capacity. This article is not intended to serve as a reference for making investment decisions.
You may also like
An AAVE whale has bought over 40,000 AAVE on dips again in the past 5 days
Chainlink Reserve adds 89,079 LINK tokens, bringing total holdings to 973,752 LINK
A certain whale has bought another $7.1 million worth of AAVE in the past 5 days
In the past 7 days, only 11 public chains had fee revenues exceeding $100,000.
