Alibaba launches more efficient Qwen3-Next AI model
Jinse Finance reported that Alibaba's Tongyi Qianwen has released the next-generation foundational model architecture, Qwen3-Next, and open-sourced the Qwen3-Next-80B-A3B series models based on this architecture. Compared to the MoE model structure of Qwen3, this architecture features the following core improvements: hybrid attention mechanism, high-sparsity MoE structure, a series of training-stable optimizations, and a multi-token prediction mechanism that enhances inference efficiency. Based on the Qwen3-Next model structure, Alibaba has trained the Qwen3-Next-80B-A3B-Base model, which has 80 billion parameters but only activates 3 billion parameters. This Base model achieves performance comparable to or slightly better than the Qwen3-32B dense model, while its training cost (GPU hours) is less than one-tenth of Qwen3-32B, and its inference throughput for contexts above 32k is more than ten times that of Qwen3-32B, achieving exceptional cost-effectiveness in both training and inference.
Disclaimer: The content of this article solely reflects the author's opinion and does not represent the platform in any capacity. This article is not intended to serve as a reference for making investment decisions.
You may also like
Today's Fear and Greed Index rises to 57, shifting from Neutral to Greed
CleanCore Solutions increases DOGE holdings to 500 millions
Elon Musk issues another warning about the $37 trillion national debt
DTCC lists three crypto ETFs, covering Solana, HBAR, and XRP
Trending news
MoreCrypto prices
More








