Back

Qwen 2 72B Instruct

Qwen
Input: text
Output: text
Released: Jun 7, 2024Updated: Mar 28, 2025

Qwen2 72B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.

It features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization.

For more details, see this blog post and GitHub repo.

Usage of this model is subject to Tongyi Qianwen LICENSE AGREEMENT.

32,768 Token Context

Process and analyze large documents and conversations.

Advanced Coding

Improved capabilities in front-end development and full-stack updates.

Agentic Workflows

Autonomously navigate multi-step processes with improved reliability.

Available On

ProviderModel IDContextMax OutputInput CostOutput CostThroughputLatency
Togethertogether33K4K$0.90/M$0.90/M36.6 t/s602 ms
Standard Pricing
Input Tokens
$0.0000009

per 1K tokens

Output Tokens
$0.0000009

per 1K tokens

Do Work. With AI.