Qwen iconQwen: Qwen3 Coder Next

Model Type

Proprietary model icon

Proprietary Model

API access only

Recommended Use Cases

Text Generation

Try Qwen3 Coder Next

Qwen3-Coder-Next is Alibaba's ultra-efficient open-weight coding model built on the Qwen3-Next architecture, featuring 80B total parameters with only 3B active per token, designed specifically for coding agents and local development.

Qwen3-Coder-Next achieves performance comparable to models with 10-20x more active parameters, making it highly cost-effective for agent deployment.

  • Qwen Team

Overview

Qwen3-Coder-Next is built on Qwen3-Next-80B-A3B-Base using a novel hybrid attention and MoE architecture. It has been agentically trained at scale on 800,000 executable tasks with environment interaction and reinforcement learning, optimizing for long-horizon reasoning, complex tool usage, and recovery from execution failures.

Key Features

  • Ultra-efficient inference: 80B total parameters, 3B active per token
  • 256K native context: Full repository-scale understanding without chunking
  • Agentic training: Trained on 800K verifiable coding tasks with executable environments
  • 370 programming languages: Expanded from 92 in previous versions
  • Tool calling: Works with Claude Code, Qwen Code, Cline, and other agent frontends
  • Non-thinking mode only: Direct responses without <think> blocks

Technical Specifications

SpecificationValue
Total Parameters80B
Active Parameters3B
ArchitectureHybrid attention + sparse MoE
Layers48 (Gated DeltaNet + Gated Attention + MoE)
Context Length256K tokens
LicenseApache 2.0

When to Use Qwen3-Coder-Next

Choose Qwen3-Coder-Next when you need:

  • Local deployment with limited GPU resources
  • Cost-effective agentic coding at scale
  • Integration with existing IDE/CLI tools
  • Repository-scale code understanding

Consider alternatives when you need:

  • Maximum capability (use Qwen3-Coder-480B-A35B)
  • Extended thinking mode for complex reasoning
  • Vision inputs for coding from mockups (use Qwen3-VL)

Availability

  • Open Weights: Hugging Face (Qwen/Qwen3-Coder-Next)
  • Inference: SGLang (>=0.5.8), vLLM (>=0.15.0)
  • Local: Ollama, LMStudio, llama.cpp

Role in Series

Qwen3 Coder models offer different capability/efficiency tradeoffs:

  1. Qwen3-Coder-Next: Most efficient, 80B/3B active, local deployment (this model)
  2. Qwen3-Coder-30B-A3B: Balanced efficiency, 30B/3B active
  3. Qwen3-Coder-480B-A35B: Maximum capability, 480B/35B active

Links