Back

Qwen 2 72B Instruct

Qwen
Input: text
Output: text
Released: Jun 7, 2024Updated: Mar 28, 2025

Qwen2 72B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.

It features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization.

For more details, see this blog post and GitHub repo.

Usage of this model is subject to Tongyi Qianwen LICENSE AGREEMENT.

32,768 Token Context

Process and analyze large documents and conversations.

Advanced Coding

Improved capabilities in front-end development and full-stack updates.

Agentic Workflows

Autonomously navigate multi-step processes with improved reliability.

Available On

ProviderModel IDContextMax OutputInput CostOutput CostThroughputLatency
Togethertogether33K4K$0.90/M$0.90/M41.5 t/s657 ms

Standard Pricing

Input Tokens

per 1M tokens

$0.90

Output Tokens

per 1M tokens

$0.90

Do Work. With AI.