Qwen: Qwen3 235B A22B

qwen/qwen3-235b-a22b

qwen/qwen3-235b-a22b-04-28
Created Apr 28, 202540,960 context
$0.13/M input tokens$0.60/M output tokens

Qwen3-235B-A22B is a 235B parameter mixture-of-experts (MoE) model developed by Qwen, activating 22B parameters per forward pass. It supports seamless switching between a "thinking" mode for complex reasoning, math, and code tasks, and a "non-thinking" mode for general conversational efficiency. The model demonstrates strong reasoning ability, multilingual support (100+ languages and dialects), advanced instruction-following, and agent tool-calling capabilities. It natively handles a 32K token context window and extends up to 131K tokens using YaRN-based scaling.

Recent activity on Qwen3 235B A22B

Tokens processed per day

Apr 28May 2May 6May 10May 14May 18May 22May 26May 30Jun 3Jun 7Jun 11Jun 15Jun 190800M1.6B2.4B3.2B

More models from Qwen

    Qwen: Qwen3 235B A22B – Recent Activity | OpenRouter