Overview
Qwen3-235B-A22B is Alibaba’s flagship open-source MoE LLM (Apache-2.0): 235B total parameters with 22B activated (128 experts, 8 active). It uniquely toggles thinking vs. non-thinking modes, supports 100+ languages, and excels at agentic tool use. Context is 32K native (≈131K with YaRN); the Instruct-2507 variant offers 256K native and up to ~1M tokens.
Description
For context length, the base A22B runs 32,768 tokens natively and ~131,072 with YaRN rope scaling. Alibaba also publishes tuned variants: Qwen3-235B-A22B-Instruct-2507 (non-thinking) with a 262K native window extendable to ~1.01M, and a Thinking-2507 line aimed at harder reasoning. Qwen3 models are supported in Hugging Face Transformers, vLLM, and SGLang, with guidance for local runtimes (Ollama, llama.cpp, MLX-LM).
About Alibaba
Chinese e-commerce and cloud leader behind Taobao, Tmall, and Alipay.