We use essential cookies to make our site work. With your consent, we may also use non-essential cookies to improve user experience and analyze website traffic…
Qwen/Qwen3-Coder-480B-A35B-Instruct cover image
featured

Qwen/Qwen3-Coder-480B-A35B-Instruct

Qwen3-Coder-480B-A35B-Instruct is the Qwen3's most agentic code model, featuring Significant Performance on Agentic Coding, Agentic Browser-Use and other foundational coding tasks, achieving results comparable to Claude Sonnet.

Qwen3-Coder-480B-A35B-Instruct is the Qwen3's most agentic code model, featuring Significant Performance on Agentic Coding, Agentic Browser-Use and other foundational coding tasks, achieving results comparable to Claude Sonnet.

Public
$0.40/$1.60 in/out Mtoken
fp8
262,144
Function
Qwen/Qwen3-Coder-480B-A35B-Instruct cover image

Qwen3-Coder-480B-A35B-Instruct

Ask me anything

0.00s

Qwen3-Coder-480B-A35B-Instruct is the Qwen3's most agentic code model to date. Qwen3-Coder is available in multiple sizes, and Qwen3-Coder-480B-A35B-Instruct is its most powerful variant, featuring the following key enhancements:

  • Significant Performance among open models on Agentic Coding, Agentic Browser-Use, and other foundational coding tasks, achieving results comparable to Claude Sonnet.
  • Long-context Capabilities with native support for 256K tokens, extendable up to 1M tokens using Yarn, optimized for repository-scale understanding.
  • Agentic Coding supporting for most platfrom such as Qwen Code, CLINE, featuring a specially designed function call format.

image/jpeg

Model Overview

Qwen3-480B-A35B-Instruct has the following features:

  • Type: Causal Language Models
  • Training Stage: Pretraining & Post-training
  • Number of Parameters: 480B in total and 35B activated
  • Number of Layers: 62
  • Number of Attention Heads (GQA): 96 for Q and 8 for KV
  • Number of Experts: 160
  • Number of Activated Experts: 8
  • Context Length: 262,144 natively.

NOTE: This model supports only non-thinking mode and does not generate **\<think>****\</think>** blocks in its output. Meanwhile, specifying enable_thinking=False is no longer required.

For more details, including benchmark evaluation, hardware requirements, and inference performance, please refer to our blog, GitHub, and Documentation.

Unlock the most affordable AI hosting

Run models at scale with our fully managed GPU infrastructure, delivering enterprise-grade uptime at the industry's best rates.