TAAFT
Free mode
100% free
Freemium
Free Trial
Create tool

Qwen3-235B-A22B

By Alibaba
New Text Gen
Released: July 25, 2025

Overview

Qwen3-235B-A22B is Alibaba’s flagship open-source MoE LLM (Apache-2.0): 235B total parameters with 22B activated (128 experts, 8 active). It uniquely toggles thinking vs. non-thinking modes, supports 100+ languages, and excels at agentic tool use. Context is 32K native (≈131K with YaRN); the Instruct-2507 variant offers 256K native and up to ~1M tokens.

Description

Alibaba’s Qwen3-235B-A22B is a flagship Mixture-of-Experts language model in the Qwen3 family. It has 235B total parameters with 22B activated per token (A22B), 128 experts with 8 active, and ships under the Apache-2.0 license. The model’s hallmark is a built-in switch between “thinking” (emits a <think>…</think> trace for deeper reasoning) and non-thinking modes, plus strong agent/tool calling and multilingual coverage across 100+ languages.
For context length, the base A22B runs 32,768 tokens natively and ~131,072 with YaRN rope scaling. Alibaba also publishes tuned variants: Qwen3-235B-A22B-Instruct-2507 (non-thinking) with a 262K native window extendable to ~1.01M, and a Thinking-2507 line aimed at harder reasoning. Qwen3 models are supported in Hugging Face Transformers, vLLM, and SGLang, with guidance for local runtimes (Ollama, llama.cpp, MLX-LM).

About Alibaba

Chinese e-commerce and cloud leader behind Taobao, Tmall, and Alipay.

Website: alibaba.com
View Company Profile

Related Models

Last updated: September 17, 2025