Back

Qwen: Qwen3 235B A22B

Qwen3
Input: text
Output: text
Released: Apr 28, 2025Updated: May 11, 2025

Qwen3-235B-A22B is a 235B parameter mixture-of-experts (MoE) model developed by Qwen, activating 22B parameters per forward pass. It supports seamless switching between a "thinking" mode for complex reasoning, math, and code tasks, and a "non-thinking" mode for general conversational efficiency. The model demonstrates strong reasoning ability, multilingual support (100+ languages and dialects), advanced instruction-following, and agent tool-calling capabilities. It natively handles a 32K token context window and extends up to 131K tokens using YaRN-based scaling.

40,960 Token Context

Process and analyze large documents and conversations.

Hybrid Reasoning

Choose between rapid responses and extended, step-by-step processing for complex tasks.

Advanced Coding

Improved capabilities in front-end development and full-stack updates.

Agentic Workflows

Autonomously navigate multi-step processes with improved reliability.

Available On

ProviderModel IDContextMax OutputInput CostOutput CostThroughputLatency
DeepInfradeepInfra41K41K$0.14/M$0.60/M25.2 t/s964 ms
kluster.aiklusterAi41K41K$0.14/M$2.00/M25.1 t/s989 ms
Parasailparasail41K41K$0.18/M$0.85/M48.0 t/s581 ms
Togethertogether41K-$0.20/M$0.60/M31.6 t/s932 ms
Nebius AI StudionebiusAiStudio41K-$0.20/M$0.60/M25.2 t/s598 ms
NovitaAInovitaAi41K41K$0.20/M$0.80/M22.8 t/s983 ms
Fireworksfireworks128K-$0.22/M$0.88/M57.5 t/s996 ms
GMICloudgmiCloud33K-$0.25/M$1.09/M44.9 t/s3103 ms
Standard Pricing
Input Tokens
$0.00000014

per 1K tokens

Output Tokens
$0.0000006

per 1K tokens

Do Work. With AI.