Qwen-MT: AI-Powered Machine Translation

Qwen-MT (Machine Translation) is Alibaba's dedicated translation model — purpose-built for high-quality multilingual translation across 92 languages covering 95%+ of the world's population. Released in July 2025, it's trained on the Qwen3 architecture with reinforcement learning specifically optimized for translation accuracy and fluency. Available in three tiers (Flash, Plus, Lite) with API pricing starting at $0.50 per million tokens, it outperforms GPT-4.1-mini and Gemini-2.5-Flash on translation benchmarks while competing with GPT-4.1 at a fraction of the cost.

Navigate this guide:

What Is Qwen-MT?

While general-purpose LLMs like Qwen 3.5 can translate text, they're not optimized for it — they can produce inconsistent terminology, miss nuances, or hallucinate content not in the source. Qwen-MT solves this by being a translation-specialized model: same Qwen3 architecture, but fine-tuned with reinforcement learning specifically for translation quality, consistency, and domain adaptation.

The result is a model that produces translations with fewer errors, more consistent terminology, and better handling of domain-specific jargon than general-purpose alternatives — while being significantly cheaper to run.

92 Languages Supported

Qwen-MT supports 92 languages covering 95%+ of the global population. Major language families include:

Model Tiers

TierStrengthSpeedBest For
qwen-mt-plusHighest quality, nuanced translationStandardLegal, medical, literary, marketing content
qwen-mt-flashBalanced quality/speedFastGeneral business, documentation, web content
qwen-mt-liteCost-efficient, high throughputFastestBulk translation, user-generated content, chat

Key Features

Terminology Intervention

Define custom glossaries and the model will respect them throughout the translation. Critical for brands, technical documentation, and regulated industries where specific terms must be translated consistently.

Domain Prompts

Specify the domain context ("medical", "legal", "gaming", "e-commerce") and the model adapts its vocabulary, register, and style accordingly. No fine-tuning needed — just a prompt parameter.

Translation Memory

Feed previously approved translations as context, and the model will maintain consistency with your existing translated corpus. Essential for large-scale localization projects.

Reinforcement Learning Optimized

Unlike general LLMs that translate as a side effect of language understanding, Qwen-MT was explicitly optimized through RL for translation metrics: faithfulness, fluency, and terminological consistency.

Benchmarks

ModelTranslation Quality (avg)Notes
qwen-mt-plusCompetitive with GPT-4.1Significantly cheaper
qwen-mt-flashOutperforms GPT-4.1-miniSimilar speed
qwen-mt-liteOutperforms Gemini-2.5-FlashHighest throughput
GPT-4.1Reference (high)Much more expensive
Gemini-2.5-ProCompetitive with qwen-mt-plusSimilar pricing

Qwen-MT excels particularly on low-resource language pairs and domain-specific content where general models tend to struggle.

API Pricing

TierInput (per 1M tokens)Output (per 1M tokens)
qwen-mt-lite$0.50$0.50
qwen-mt-flash$1.00$1.00
qwen-mt-plus$3.00$3.00

For comparison, GPT-4.1 costs $2.00/$8.00 per 1M tokens. At the Lite tier, Qwen-MT is 4-16x cheaper than comparable GPT-4 translation.

API Quick Start

Qwen-MT uses the DashScope API with an OpenAI-compatible format:

import openai

client = openai.OpenAI(
    base_url="https://dashscope-intl.aliyuncs.com/compatible-mode/v1",
    api_key="your-dashscope-api-key"
)

response = client.chat.completions.create(
    model="qwen-mt-plus",
    messages=[
        {"role": "system", "content": "You are a professional translator. Domain: legal."},
        {"role": "user", "content": "Translate to Spanish: The party of the first part hereby agrees..."}
    ]
)
print(response.choices[0].message.content)

With Terminology Intervention

# Add a glossary parameter
response = client.chat.completions.create(
    model="qwen-mt-plus",
    messages=[
        {"role": "system", "content": "Translate EN→DE. Glossary: 'cloud computing'='Cloud-Computing', 'machine learning'='maschinelles Lernen'"},
        {"role": "user", "content": "Our cloud computing platform uses machine learning for optimization."}
    ]
)

Use Cases

Qwen-MT vs Traditional Translation

FeatureQwen-MTGoogle Translate APIDeepL API
Languages9213333
Quality (complex text)High (LLM-based)MediumHigh
Domain promptsYesNoLimited
Terminology controlYes (glossaries)LimitedYes
Translation memoryYesNoNo
Context understandingFull document contextSentence-levelParagraph-level
Pricing (per 1M chars)From ~$0.50$20.00$25.00

The key advantage of Qwen-MT over traditional NMT (Neural Machine Translation) services is full document-level context: it reads the entire document before translating, maintaining consistency of pronouns, references, and terminology throughout — something sentence-level translation APIs struggle with.

Limitations

FAQ

Is Qwen-MT a separate product from Qwen Chat?

Qwen-MT is available both as a standalone API (via DashScope) and as a feature within Qwen Chat. The API gives you programmatic access with glossaries and domain prompts; Qwen Chat offers a simpler interface for one-off translations.

Can I use it for real-time chat translation?

Yes. The Lite and Flash tiers are fast enough for real-time chat scenarios. For voice-to-voice translation, pair it with Qwen3-ASR for input and Qwen3-TTS for output.

How does it compare to just asking Qwen 3.5 to translate?

Qwen 3.5 can translate, but Qwen-MT is specifically optimized for translation: better consistency, lower hallucination rate, glossary support, and 2-5x cheaper per token for translation workloads.

Is the model open-source?

No. Qwen-MT is currently available as an API service only. The underlying Qwen3 models are open-source, but the translation-specific fine-tuning is proprietary.