Qwen iconQwen: Qwen3 VL 235B A22B Thinking

Model Type

Open weight model icon

Open Weight Model

235B parameters

Recommended Use Cases

Text Generation

Qwen3-VL-235B-A22B-Thinking is Alibaba's most powerful reasoning-enhanced vision-language model, with 235B total parameters and 22B active per token, optimized for complex multimodal reasoning and STEM tasks.

Overview

Qwen3-VL-235B-A22B-Thinking is the reasoning-optimized variant of the flagship Qwen3-VL model. It combines massive MoE capacity with extended chain-of-thought capabilities, excelling at mathematical reasoning from diagrams, scientific visual analysis, and multi-step video understanding where accuracy outweighs speed.

Key Features

  • Flagship reasoning: Maximum capability with deliberate thinking
  • Extended thinking traces: Visible <think> blocks for complex problems
  • 1M context: 256K native, expandable to 1M tokens
  • STEM excellence: Mathematical and scientific reasoning from visuals
  • 100% AIME25: Top benchmark performance on math reasoning

Technical Specifications

SpecificationValue
Total Parameters235B
Active Parameters22B
ArchitectureMoE transformer with CoT training
Vision EncoderDeepStack multi-level ViT fusion
Context Length256K tokens (expandable to 1M)
Release DateSeptember 2025

Thinking vs Instruct (235B-A22B)

AspectThinking (this model)Instruct
Response styleExtended chain-of-thoughtDirect answers
LatencyHigherLower
Token consumptionHigherLower
Accuracy on hard tasksMaximumHigh
Best forSTEM, complex reasoningProduction, general tasks

When to Use Qwen3-VL-235B-A22B-Thinking

Choose this model when you need:

  • Maximum accuracy on complex visual reasoning
  • Mathematical problem solving from diagrams
  • Scientific analysis requiring step-by-step logic
  • Long video understanding with causal reasoning
  • Research and analysis where accuracy is critical

Consider alternatives when you need:

  • Fast production responses (use Instruct variant)
  • Lower deployment costs (use 30B-A3B models)
  • Edge deployment (use 8B models)

Availability

  • Open Weights: Hugging Face (Qwen/Qwen3-VL-235B-A22B-Thinking)
  • API: OpenRouter, DeepInfra, Novita
  • Local: vLLM with tensor parallelism (~471GB weights)

Role in Series

Qwen3-VL models by reasoning capability:

  1. Qwen3-VL-8B-Instruct: Fast, edge deployment
  2. Qwen3-VL-8B-Thinking: Edge reasoning
  3. Qwen3-VL-30B-A3B-Instruct: Efficient production
  4. Qwen3-VL-30B-A3B-Thinking: Efficient reasoning
  5. Qwen3-VL-235B-A22B-Instruct: Flagship production
  6. Qwen3-VL-235B-A22B-Thinking: Maximum reasoning (this model)

Links