Gemini API Pricing Calculator

Estimate your monthly costs for Google's Gemini models, including Gemini 1.5 Flash and Pro. Account for input tokens, output tokens, and context length.

Model Selection

Usage Parameters

Estimated Monthly Cost

Input Cost$0.00
Output Cost$0.00
Total Cost$0.00

*Estimates based on current Gemini API pricing. Actual costs may vary.

FAQ

Gemini API Pricing FAQs

Got questions? We've got answers. Here are the most common questions we get from potential clients.

Flash is optimized for speed and cost-efficiency, making it ideal for high-volume tasks. Pro is a larger model designed for complex reasoning and handling massive context windows (up to 2M tokens).

Complete Guide to Gemini API Costs

Google's Gemini API offers a flexible pricing model based on token usage. The cost depends on the model you choose (e.g., the cost-effective Flash or the powerful Pro) and the length of your context window. This is particularly relevant for businesses using AI Voice Agents that process large amounts of audio data.

Key Takeaways

  • Gemini 1.5 Flash is optimized for speed and is significantly cheaper than GPT-4o mini.
  • Gemini 1.5 Pro offers a massive 2M token context window for deep analysis.
  • Context Caching can reduce input costs by ~90% for repeated large prompts.

Gemini API Pricing Overview

ModelInput Price ($/1M)Output Price ($/1M)Cached Input ($/1M)
Multimodal Models (Text, Image, Video, Audio)
Gemini 2.5 Flash$0.15$0.60-
Gemini 2.5 Pro$1.25$10.00-
Gemini 1.5 Flash$0.075$0.30$0.018
Gemini 1.5 Pro$1.25$5.00$0.31
Embedding Models
Text Embedding 004$0.025$--
Multimodal Embedding 001$0.0002 / image$--

Gemini 1.5 Pro: The Powerhouse

Gemini 1.5 Pro offers a massive 2 million token context window, making it perfect for analyzing large documents, codebases, or long videos. It's a key component in process optimization tasks.

Gemini 1.5 Flash: Speed and Efficiency

Gemini 1.5 Flash is designed for high-frequency, low-latency tasks. It's incredibly affordable and fast, suitable for real-time applications.

Key Pricing Factors

  • Model Selection: Gemini 1.5 Flash is significantly cheaper than 1.5 Pro, making it ideal for high-volume tasks.
  • Context Window: While Gemini 1.5 Pro has a 2M token window, prompts longer than 128k tokens may incur higher costs in some legacy tiers, though current pricing is generally flat.
  • Input vs. Output: Output tokens are significantly more expensive than input tokens. Optimizing your prompts to generate concise responses can save money.

How to Optimize Costs

To keep your Gemini API bills low, consider using Context Caching for repeated content, which offers a significant discount (up to 75% cheaper for cached inputs). Also, use the Flash model for simpler tasks where the reasoning capabilities of Pro aren't strictly necessary.