Gemini 2.0 Flash
Budgetby Google
Gemini 2.0 Flash is Google's speed-optimized model, delivering remarkably fast inference at just $0.10 per million input tokens. It retains the 1M token context window, making it one of the most cost-effective options for high-volume workloads.
Input Price
$0.10
per 1M tokens
Output Price
$0.40
per 1M tokens
Context Window
1M
tokens
Released
2025-02
API access
Capabilities
textvisiontool-usecode
Key Strengths
- ✓Ultra-low cost ($0.10/1M input)
- ✓1M token context
- ✓Fastest in class
- ✓Full multimodal support
Best For
- ▸High-volume API workloads
- ▸Real-time applications
- ▸Chat and classification
- ▸Batch processing
Benchmark Scores
| Benchmark | Score | Description |
|---|---|---|
| MMLU-Pro | 84.5 | General knowledge and reasoning across 57 subjects |
| HumanEval | 87.6 | Python code generation and problem solving |
| GPQA Diamond | 54.8 | Graduate-level science questions verified by domain experts |
| MATH | 77.2 | Competition-level mathematics problems |
| SWE-bench | 43.1 | Real-world software engineering tasks from GitHub issues |
Scores sourced from public benchmark datasets. See full benchmark leaderboard for all models.
Pricing Details
Input tokens
$0.10
per 1M tokens
Output tokens
$0.40
per 1M tokens
Estimated cost per 1K requests
$0.30
~1K input + ~500 output tokens avg
Prices are subject to change. Check the official documentation for current pricing. See the cost calculator for detailed estimates.