Capabilities, modalities, and lifecycle fields pulled from the model database.
Comparative results across benchmarks shared by the selected models.
| SimpleQA | 50.8% |
| Humanity's Last Exam | 17.8% |
| MMLU | 88.6% |
| Aider-Polyglot | 72.7% |
| AIME 2025 | 83.0% |
| SWE-Bench | 63.2% |
| Thematic Generalisation | 175.0% |
| MMMU | 79.6% |
| Confabulations | 1062.0% |
| EQ-Bench 3 | 124700.0% |
| GPQA Diamond | 83.0% |
| LMArena Text | 144600.0% |
| NYT Connections | 42.5% |
Observed provider pricing per million tokens.
No providers found yet.
Usage and distribution terms.
Model release chronology.
A deeper field-by-field view (including benchmarks, pricing, and links).
| General Information | |
| Context Window | Input: - Output: - |
| Modalities | In: - Out: - |
| Reasoning | - |
| Web access | - |
| Parameters | - |
| Training Tokens | - |
| License | Proprietary |
| Knowledge Cutoff | - |
| Status | Retired |
| Release | May 2025 |
| Announced | May 2025 |
| Deprecation | Jun 2025 |
| Retirement | Jul 2025 |
| Links | - |
| Operational Metrics | |
| Cost per 1M Tokens | Input: - Output: - |
| Latency | - |
| Throughput | - |
| Benchmarks | |
| AIME 2025 | |
| Aider-Polyglot | |
| Confabulations | |
| EQ-Bench 3 | |
| GPQA Diamond | |
| Humanity's Last Exam | |
| LMArena Text | |
| MMLU | |
| MMMU | |
| NYT Connections | |
| SWE-Bench | |
| SimpleQA | |
| Thematic Generalisation | |