Capabilities, modalities, and lifecycle fields pulled from the model database.
Comparative results across benchmarks shared by the selected models.
| Humanity's Last Exam | 44.4% |
| Tau 2 Retail | 90.8% |
| SWE-Bench | 80.6% |
| SWE Bench Pro | 54.2% |
| LiveCodeBench Pro | 288700.0% |
| SciCode | 59.0% |
| APEX-Agents | 33.5% |
| GDPval-AA | 131700.0% |
| Tau 2 Telecom | 99.3% |
| Scale MCP Atlas | 69.2% |
| BrowseComp | 85.9% |
| MMMU Pro | 80.5% |
| MMMLU | 92.6% |
| OpenAI MRCR 8 Needle 128k | 84.9% |
| OpenAI MRCR 8 Needle 1m | 26.3% |
| ARC-AGI-1 | 98.0% |
| ARC-AGI-2 | 77.1% |
| GPQA Diamond | 94.3% |
| Terminal Bench 2.0 | 68.5% |
Observed provider pricing per million tokens.
All unique meters observed across the selected models.
| Meter | Gemini 3.1 Pro Preview |
|---|---|
| Input Text Tokens | $2.00 |
Providers that expose each model based on observed pricing data.
Maximum input and output token capacity.
Usage and distribution terms.
Model release chronology.
Most recent training data date (when available).
A deeper field-by-field view (including benchmarks, pricing, and links).
| General Information | |
| Context Window | Input: 1,048,576 Output: 65,536 |
| Modalities | In: Text, Vision, Video, Audio Out: Text |
| Reasoning | - |
| Web access | - |
| Parameters | - |
| Training Tokens | - |
| License | Proprietary |
| Knowledge Cutoff | Jan 2025 |
| Status | Available |
| Release | Feb 2026 |
| Announced | Feb 2026 |
| Deprecation | - |
| Retirement | - |
| Links | |
| Operational Metrics | |
| Cost per 1M Tokens | Input: $2.00 Output: $12.00 |
| Latency | - |
| Throughput | - |
| Benchmarks | |
| APEX-Agents | |
| ARC-AGI-1 | |
| ARC-AGI-2 | |
| BrowseComp | |
| GDPval-AA | |
| GPQA Diamond | |
| Humanity's Last Exam | |
| LiveCodeBench Pro | |
| MMMLU | |
| MMMU Pro | |
| OpenAI MRCR 8 Needle 128k | |
| OpenAI MRCR 8 Needle 1m | |
| SWE Bench Pro | |
| SWE-Bench | |
| Scale MCP Atlas | |
| SciCode | |
| Tau 2 Retail | |
| Tau 2 Telecom | |
| Terminal Bench 2.0 | |