Qwen 3.5 Debuts with Massive Reinforcement Learning Upgrades and Native Multimodality.

Qwen 3.5 Debuts with Massive Reinforcement Learning Upgrades and Native Multimodality.
Alibaba Qwen 3.5 Debuts: Massive 1-Million Token Context and Disruptive Pricing

The Qwen team at Alibaba has officially launched Qwen 3.5, the latest generation of its Large Language Model (LLM). This new iteration introduces native multimodal capabilities, allowing the model to process image inputs directly. Furthermore, the proprietary version of the model now supports a massive 1-million token context window, all while significantly slashing operational costs.

The Efficiency of MoE Architecture

The flagship open-weight release, Qwen 3.5-397B-A17B, showcases a remarkable feat of engineering. Although the model houses a total of 397 billion parameters, its Mixture-of-Experts (MoE) architecture ensures that only 17 billion parameters are activated during inference.

  • Advanced Training: The model was refined through intensive Reinforcement Learning (RL), with expanded task categories specifically designed for RL-based optimization.

  • Open Model Capacity: This version supports an input capacity of up to 263,100 tokens.

Disruptive Pricing and the "Plus" Variant

Alibaba is positioning Qwen 3.5 as a high-performance yet affordable solution. The Qwen 3.5-397B-A17B is priced at:

  • Input: $0.6 per million tokens

  • Output: $3.6 per million tokens

Additionally, the team introduced the proprietary Qwen 3.5 Plus (2026-02-15). The standout feature of this closed-source model is its ability to handle a 1-million token context, placing it in direct competition with top-tier global models for long-form data processing.

The 397B-A17B architecture signifies a model with a massive "brain" but intelligent enough to selectively select only a few relevant "experts" (17B) to answer questions. This saves enormous computing power while maintaining the knowledge level of a 400B model.

Alibaba's strong emphasis on RL in this version reflects the 2026 trend where models improve not just through "reading a lot" (pre-training), but through "reasoning" via trial and error. This makes Qwen 3.5 exceptionally strong in coding and mathematics tasks.

Unlike previous versions that often simply "pasted" visual models onto language models, Qwen 3.5 uses Native Visual Processing, meaning it understands the connections between images and text more deeply. Examples include analyzing complex stock charts or understanding engineering diagrams.

The increase to 1 million tokens in the Plus version is a direct declaration of war on Gemini 1.5 Pro and Claude 3.5, allowing users to "throw" entire documents into it. Or, you can input the entire project's codebase so the AI ​​can summarize or continue working immediately. 

 

 

Inside the Crypto Powered Markets of Modern Slavery.

 

Source: Qwen 

Comments

Popular posts from this blog

DavaIndia Data Breach How a Simple Misconfiguration Exposed 17,000 Medical Orders.

Airbnb’s 2026 AI Evolution Focuses on Personal Assistants and Voice Support.

OpenAI Sunsets GPT-4o Ending the Era of "Sycophantic AI" for Public Safety.

When AI Agent Attack A New Era of Harassment in the Open Source Community.

Lenovo Breaks Records with $22.2 Billion Quarter as AI Portfolio Hits 72% Growth.

Google Docs Unveils "Audio Summaries" Let Gemini Turn Your Documents into Mini-Podcasts.

Notepad No Longer "Safe"? Microsoft Patches Remote Code Execution Flaw in February Update.