"The Qwen3.5 Native Vision-Language Series Plus model is built on a hybrid architecture that integrates linear attention mechanisms with sparse Mixture-of-Experts (MoE), achieving significantly higher inference efficiency. Across various task evaluations, the 3.5 series demonstrates exceptional performance comparable to current state-of-the-art frontier models. Compared to the Qwen 3 series, this model represents a massive leap forward in both text-only and multimodal capabilities."'
EDIT: Okay, there's some limited information at https://modelstudio.console.alibabacloud.com/ap-southeast-1/...