Alibaba Launches Qwen3.5 with Native Multimodal and Reasoning Capabilities
- •Alibaba releases Qwen3.5-397B-A17B, ranking third among global open weights models.
- •New unified architecture supports native vision, video, and reasoning modes in one model.
- •Model achieves massive gains in agentic tasks while maintaining high parameter efficiency.
Alibaba has officially entered the next generation of open-source AI with the debut of Qwen3.5-397B-A17B, a model that marks a significant architectural shift for the tech giant. Unlike its predecessors, which often split text and vision capabilities into separate versions, this new release follows the industry trend toward "native multimodality." This means the model processes images and videos directly within its core structure rather than relying on external plugins or secondary systems.
The model’s efficiency is particularly striking; despite having nearly 400 billion total parameters, it only uses 17 billion active parameters during any single calculation, using a Mixture-of-Experts architecture. This lean approach allows it to punch well above its weight class, outperforming much larger rivals in complex tasks involving Agentic AI—systems that act as digital assistants by planning and executing multi-step workflows like data analysis or coding.
However, independent benchmarking reveals that while its intelligence has surged, the model still struggles with accuracy relative to its top-tier peers. While it performs significantly better than previous versions, it still lags behind competitors in knowing when to admit it doesn't have an answer, a behavior often termed a hallucination. For developers, the release under an Apache 2.0 license represents a massive win for the open-weights ecosystem, providing a top-tier tool for specialized, local deployment.