Qwen AI

Alibaba Cloud's Open-Source AI

The most comprehensive ecosystem of open-weight language, vision, audio and multimodal models — from the 0.6B edge model to the 397B flagship.

100+
Open Models
40M+
Downloads
201
Languages
Apache 2.0
License

Explore the Qwen Ecosystem

Flagship & General Purpose

Qwen 3.5
397B MoE flagship — vision, language, reasoning, 1M context
Local + API8 sizes
Qwen 3
0.6B to 235B + Qwen3-Next hybrid architecture
Local + API
Qwen Max
Most powerful API model — Max-Thinking + 3.5-Max-Preview
API only
Qwen Plus
Cost-effective API model for production workloads
API only

Coding & Reasoning

Qwen Coder
70.6% SWE-Bench with 3B active params — Coder-Next + CLI
Local + API
QwQ
Reasoning specialist — math, logic, and analysis
Local
Qwen Math
Specialized mathematical reasoning and problem-solving
Local + API
Qwen Code
Free open-source CLI coding agent — 18K GitHub stars
LocalFree

Vision, Image & Search

Qwen Vision
2-hour video understanding, OCR in 32 languages, visual agents
Local + API6 sizes
Qwen Image
AI image generation and editing — Image 2.0 (7B)
Local + API
Qwen Embeddings
#1 MTEB multilingual — text + multimodal embeddings & rerankers
Local + API

Audio & Voice

Qwen Omni
All-in-one: text, image, audio, video in, speech out
Local + API
Qwen Audio
Audio understanding and analysis
Local + API
Qwen ASR
Speech recognition in 52+ languages
Local + API
Qwen TTS
Text-to-speech with voice cloning and design
Local + API
Qwen Translate
Machine translation for 92 languages via API
API only

Guides & Resources

Can I Run Qwen?
Check which models run on your hardware — with direct GGUF downloads
Run Locally (Hub)
Compare Ollama, llama.cpp, LM Studio, and MLX
Ollama Guide
Easiest setup — one command, known issues documented
llama.cpp Guide
Maximum speed — 2-5x faster than Ollama on NVIDIA
LM Studio Guide
Visual interface — download, chat, serve API with a GUI
MLX Guide (Mac)
Fastest on Apple Silicon — 2x speed, 50% less memory
Use Cases
Coding, research, business — find the right model
Prompt Engineering
Templates and best practices for great results

How Does Qwen Compare?

See how Qwen stacks up against other leading AI platforms.

Qwen vs DeepSeek
Qwen 3.5 vs DeepSeek V3 — benchmarks, ecosystem, multimodal
Qwen vs ChatGPT
Open-source vs proprietary — features, pricing, capabilities
Qwen vs Mistral
Two leading open-weight families compared

What's New in the Qwen Ecosystem

Mar 2026
Qwen3.5-Max-Preview enters LMArena Top 10
Alibaba's closed-weight flagship debuts at rank 10 in Expert Prompts, surpassing GPT-5.4 and Claude Opus 4.5.
Mar 2026
Qwen 3.5 small models: 0.8B to 9B
Four new dense models complete the lineup. The 9B beats OpenAI's 120B model on MMLU-Pro and GPQA Diamond.
Feb 2026
Qwen 3.5 launches: 397B-A17B flagship
Hybrid GatedDeltaNet architecture, native multimodal, 201 languages, 262K context. 8 open-weight models under Apache 2.0.
Feb 2026
Qwen3-Coder-Next: 70.6% SWE-Bench with 3B active
Ultra-sparse coding agent trained with RL from environment feedback. Matches Claude Sonnet 4.0-level coding on consumer hardware.
Jan 2026
Qwen3-Max-Thinking released
Test-time scaling with multi-pass reasoning. 10-12x cheaper than GPT-5.2 via API with competitive performance.

What Is Qwen AI?

Qwen (pronounced "chwen", from the Chinese 通义千问 / Tongyì Qianwèn) is a family of large language models and multimodal AI systems developed by Alibaba Cloud. Since its initial release in 2023, the project has grown into one of the largest open-source AI ecosystems in the world, spanning text generation, coding, vision, audio, speech, image generation, and real-time multimodal interaction.

The latest flagship, Qwen 3.5 (February 2026), is a 397B Mixture-of-Experts model with 17B active parameters, 262K native context extensible to 1M tokens, and native support for 201 languages. It competes directly with GPT-5.2, Claude Opus 4.5, and Gemini 3 Pro across reasoning, coding, and multimodal tasks — while remaining open-weight under the Apache 2.0 license.

Most Qwen models can be run locally with tools like Ollama, llama.cpp, or LM Studio, deployed via Alibaba Cloud's DashScope API, or accessed through Qwen Chat for free. The project is backed by Alibaba Cloud's infrastructure, serving 290,000+ enterprise customers worldwide. Looking for previous-generation Qwen 2.5 models?

Start Building with Qwen AI

Open-source models under Apache 2.0, managed APIs, and comprehensive guides. From prototype to production in days.
Qwen 3.5 Overview Can I Run Qwen? Run Locally

Guides and benchmarks updated March 2026.