Gemini 2.0 Flash

Budget

by Google

Gemini 2.0 Flash is Google's speed-optimized model, delivering remarkably fast inference at just $0.10 per million input tokens. It retains the 1M token context window, making it one of the most cost-effective options for high-volume workloads.

Input Price

$0.10

per 1M tokens

Output Price

$0.40

per 1M tokens

Context Window

1M

tokens

Released

2025-02

API access

Capabilities

textvisiontool-usecode

Key Strengths

  • Ultra-low cost ($0.10/1M input)
  • 1M token context
  • Fastest in class
  • Full multimodal support

Best For

  • High-volume API workloads
  • Real-time applications
  • Chat and classification
  • Batch processing

Benchmark Scores

BenchmarkScoreDescription
MMLU-Pro84.5General knowledge and reasoning across 57 subjects
HumanEval87.6Python code generation and problem solving
GPQA Diamond54.8Graduate-level science questions verified by domain experts
MATH77.2Competition-level mathematics problems
SWE-bench43.1Real-world software engineering tasks from GitHub issues

Scores sourced from public benchmark datasets. See full benchmark leaderboard for all models.

Pricing Details

Input tokens

$0.10

per 1M tokens

Output tokens

$0.40

per 1M tokens

Estimated cost per 1K requests

$0.30

~1K input + ~500 output tokens avg

Prices are subject to change. Check the official documentation for current pricing. See the cost calculator for detailed estimates.

Related Models

View DocumentationCompare ModelsCost CalculatorFull Pricing Guide