Back
Qwen 2 72B Instruct
Qwen
Input: text
Output: text
Released: Jun 7, 2024•Updated: Mar 28, 2025
Qwen2 72B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.
It features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization.
For more details, see this blog post and GitHub repo.
Usage of this model is subject to Tongyi Qianwen LICENSE AGREEMENT.
32,768 Token Context
Process and analyze large documents and conversations.
Advanced Coding
Improved capabilities in front-end development and full-stack updates.
Agentic Workflows
Autonomously navigate multi-step processes with improved reliability.
Available On
Provider | Model ID | Context | Max Output | Input Cost | Output Cost | Throughput | Latency |
---|---|---|---|---|---|---|---|
Together | together | 33K | 4K | $0.90/M | $0.90/M | 36.6 t/s | 602 ms |
Standard Pricing
Input Tokens
$0.0000009
per 1K tokens
Output Tokens
$0.0000009
per 1K tokens