This site is independently operated and is not affiliated with Google or Alphabet Inc. Verify pricing on Google's official website.

Gemini 2.5 Flash Pricing

Best Value

The sweet spot of Google's model lineup. Gemini 2.5 Flash delivers strong reasoning and generation quality at a fraction of Pro's cost, with the same 1M token context window. The model most teams should default to for customer-facing applications and production workloads.

Gemini 2.5 Flash API Pricing

Under 200K context

Input

$0.15 / MTok

Output

$0.60 / MTok

Over 200K context

Input

$0.30 / MTok

Output

$1.20 / MTok

Context window: 1M tokens
Free tier: 1,500 req/day
Context caching supported

Cost Advantage Over Pro

Flash is dramatically cheaper than Pro across both input and output. For most production workloads, the quality difference is negligible while the cost savings are enormous.

8.3x

Cheaper on input

$0.15 vs $1.25 per MTok

16.7x

Cheaper on output

$0.60 vs $10.00 per MTok

Ideal Use Cases for 2.5 Flash

Flash excels at tasks where speed and cost matter more than peak reasoning depth. For high-volume pipelines where you are processing thousands or millions of requests, the cost savings compound quickly.

Text classification

Categorise support tickets, sort emails, tag content, detect intent. Flash handles these reliably at massive scale for pennies per thousand requests.

Summarisation

Condense articles, meeting notes, reports, and documents. Flash produces clean, accurate summaries that are difficult to distinguish from Pro's output.

Data extraction

Pull structured data from unstructured text. Names, dates, addresses, product details, financial figures. Flash's extraction accuracy is excellent.

High-volume processing

Any pipeline running 10K+ requests per day. Content moderation, sentiment analysis, translation, and batch processing of customer data.

Chatbots and assistants

Customer-facing conversational AI where response speed and cost per conversation matter. Flash's lower latency is an additional advantage here.

Code assistance

Autocomplete, simple refactoring, documentation generation, and code explanation. For complex architecture decisions, consider upgrading to Pro.

High-Volume Cost Modelling

What does Flash cost at scale? Below are daily cost estimates for a typical workload: 1,000 input tokens and 500 output tokens per request (a standard chatbot or classification task).

Daily VolumeInput CostOutput CostDaily TotalMonthly Est.
10,000 req/day$1.50$3.00$4.50$135
100,000 req/day$15.00$30.00$45.00$1,350
1,000,000 req/day$150.00$300.00$450.00$13,500

Based on 1,000 input tokens and 500 output tokens per request. Monthly estimates assume 30 days. Your actual costs will vary based on prompt length and response size. Use our cost calculator for custom estimates.

What would this cost on Pro?

The same workloads on Gemini 2.5 Pro would cost dramatically more. Here is the comparison at 100,000 requests per day:

2.5 Flash (100K req/day)

$45.00 / day

2.5 Pro (100K req/day)

$625.00 / day

Input: 100K x 1K/1M x $1.25 = $125, Output: 100K x 500/1M x $10 = $500

Flash saves $580/day ($17,400/month) at this volume. That is a 92.8% cost reduction.

Flash vs Claude Haiku vs GPT-4o mini

Gemini 2.5 Flash competes directly with Claude Haiku 3.5 and GPT-4o mini as the "fast and affordable" tier from each provider. Here is how they compare on price, context window, and features.

FeatureGemini 2.5 FlashClaude Haiku 3.5GPT-4o mini
Input / MTok$0.15$0.80$0.15
Output / MTok$0.60$4.00$0.60
Context window1M tokens200K tokens128K tokens
Free tier1,500 req/dayNoneLimited
SpeedFastFastFast
Best forVolume + long contextQuick tasksVolume workloads

Flash vs Claude Haiku 3.5

Flash is 5.3x cheaper on input ($0.15 vs $0.80) and 6.7x cheaper on output ($0.60 vs $4.00). The cost difference is enormous. At 100,000 requests per day (1K in / 500 out per request), Flash costs $45/day while Haiku costs $280/day. That is $7,050/month in savings. Flash also offers a 5x larger context window (1M vs 200K tokens) and a free tier that Haiku lacks.

Flash vs GPT-4o mini

These two models are essentially the same price: $0.15/$0.60 per million tokens for both. The decision comes down to ecosystem and features. Flash offers a 1M token context window (nearly 8x larger than GPT-4o mini's 128K), a more generous free tier, and native integration with Google Cloud services. GPT-4o mini integrates with OpenAI's ecosystem including function calling and the Assistants API.

What Is Gemini 2.5 Flash?

Gemini 2.5 Flash is Google's efficiency-focused model, designed to deliver strong performance at the lowest possible cost and latency. It sits between the budget 2.0 Flash and the flagship 2.5 Pro in Google's model lineup, offering a balance of capability and affordability that makes it the recommended default for most production workloads.

Unlike Pro, which is optimised for peak reasoning quality, Flash is optimised for throughput. It processes requests faster, uses fewer compute resources, and passes those savings on through lower pricing. For many tasks, including summarisation, classification, translation, and content generation, the quality gap between Flash and Pro is difficult to measure.

Flash shares the same 1M token context window as Pro, so you do not sacrifice context capacity by choosing the cheaper model. It also supports the same multi-modal inputs (text, images, audio, video) and context caching features. The primary trade-off is in complex multi-step reasoning, where Pro demonstrates measurably better performance on benchmarks and in practice.

Frequently Asked Questions

How much does Gemini 2.5 Flash cost?
Gemini 2.5 Flash costs $0.15 per million input tokens and $0.60 per million output tokens for requests under 200K context. Over 200K context, input doubles to $0.30 and output doubles to $1.20 per million tokens. It supports a 1M token context window.
How much cheaper is Flash than Pro?
Flash is 8.3x cheaper on input ($0.15 vs $1.25 per MTok) and 16.7x cheaper on output ($0.60 vs $10.00). For a typical 100K request/day workload, Flash costs $45/day vs Pro at $625/day, saving 92.8%.
Is Gemini 2.5 Flash cheaper than Claude Haiku?
Yes. Flash costs $0.15/$0.60 vs Claude Haiku 3.5 at $0.80/$4.00 per million tokens. Flash is 5.3x cheaper on input and 6.7x cheaper on output, with a larger context window and a free tier.
How does Flash compare to GPT-4o mini?
They are priced identically at $0.15/$0.60 per million tokens. Flash offers a larger context window (1M vs 128K tokens) and a more generous free tier. GPT-4o mini integrates with OpenAI's ecosystem. Quality is comparable for most tasks.
Does Gemini 2.5 Flash have a free tier?
Yes. Google AI Studio provides 1,500 free requests per day and 15 requests per minute for Flash models. No credit card required. This covers prototyping and many low-volume production use cases.