Capabilities, modalities, and lifecycle fields pulled from the model database.
Comparative results across benchmarks shared by the selected models.
| Tau 2 Telecom | 98.2% |
| ARC-AGI-2 | 7.8% |
| ARC-AGI-1 | 35.2% |
| MMMLU | 90.8% |
| Terminal Bench 2.0 | 57.8% |
| SWE-Bench | 80.6% |
| Vending Bench 2 | 496706.0% |
| Humanity's Last Exam | 30.8% |
| GPQA Diamond | 87.0% |
| MMMU | 80.7% |
| AIME 2025 | 92.8% |
| SWE Bench Pro | 51.6% |
| Tau 2 Airline | 70.1% |
| OSWorld | 66.3% |
| SWE Bench Multilingual | 76.2% |
| Tau 2 Retail | 88.9% |
Observed provider pricing per million tokens.
All unique meters observed across the selected models.
| Meter | Claude Opus 4.5 |
|---|---|
| Input Text Tokens | $5.00 |
Providers that expose each model based on observed pricing data.
Plans that include each selected model, grouped by organisation.
3 plans
Usage and distribution terms.
Model release chronology.
A deeper field-by-field view (including benchmarks, pricing, and links).
| General Information | |
| Context Window | Input: - Output: - |
| Modalities | In: Text, Vision Out: Text |
| Reasoning | - |
| Web access | - |
| Parameters | - |
| Training Tokens | - |
| License | Proprietary |
| Knowledge Cutoff | - |
| Status | Available |
| Release | Nov 2025 |
| Announced | Nov 2025 |
| Deprecation | - |
| Retirement | - |
| Links | |
| Operational Metrics | |
| Cost per 1M Tokens | Input: $5.00 Output: $25.00 |
| Latency | - |
| Throughput | - |
| Benchmarks | |
| AIME 2025 | |
| ARC-AGI-1 | |
| ARC-AGI-2 | |
| GPQA Diamond | |
| Humanity's Last Exam | |
| MMMLU | |
| MMMU | |
| OSWorld | |
| SWE Bench Multilingual | |
| SWE Bench Pro | |
| SWE-Bench | |
| Tau 2 Airline | |
| Tau 2 Retail | |
| Tau 2 Telecom | |
| Terminal Bench 2.0 | |
| Vending Bench 2 | |