Google Gemini provides multimodal AI capabilities:
- Advanced reasoning with Gemini 2.5 and 3.0
- Multimodal understanding (text, image, audio)
- Prompt caching for cost optimization
- Cost-effective options with Flash variants
Available Models
Gemini 3 Pro
Next-generation model with enhanced capabilities and prompt caching.| Model | Input | Output | Cache Read | Best For |
|---|---|---|---|---|
| Gemini 3 Pro Preview | $2.60/M | $15.60/M | $0.26/M | Cutting-edge performance |
Gemini 3 Pro supports prompt caching - reuse context at 90% discount.
Gemini 2.5 Pro
Most capable current-gen models for complex tasks.| Model | Input | Output | Best For |
|---|---|---|---|
| Gemini 2.5 Pro | $1.62/M | $13.00/M | Production apps, complex reasoning |
| Gemini 2.5 Pro Preview (06-05) | $1.62/M | $13.00/M | Latest preview features |
| Gemini 2.5 Pro Preview (05-06) | $1.62/M | $13.00/M | Stable preview |
| Gemini 2.5 Pro Preview (03-25) | $1.62/M | $13.00/M | Earlier preview |
| Gemini 2.5 Pro Exp (03-25) | $1.62/M | $13.00/M | Experimental features |
Gemini 2.5 Flash
Fast, cost-efficient models for high-volume tasks.| Model | Input | Output | Best For |
|---|---|---|---|
| Gemini 2.5 Flash | $0.20/M | $3.25/M | Fast inference, production |
| Gemini 2.5 Flash Preview (05-20) | $0.20/M | $0.78/M | Latest features |
| Gemini 2.5 Flash Preview (04-17) | $0.20/M | $0.78/M | Standard preview |
| Gemini 2.5 Flash Preview (04-17 Thinking) | $0.20/M | $4.55/M | Extended reasoning mode |
| Gemini 2.5 Flash Lite Preview (06-17) | $0.13/M | $0.52/M | Ultra-lightweight |
Best Practices
Model Selection
Use Gemini 3 for
- Cutting-edge features
- Latest capabilities
- Prompt caching needs
- Advanced reasoning
Use 2.5 Pro for
- Complex reasoning tasks
- Production applications
- High-quality outputs
- When accuracy matters
Use Flash for
- High-volume tasks
- Fast responses needed
- Cost-sensitive workloads
- Simple queries
Prompt Caching (Gemini 3)
Optimize costs with prompt caching on Gemini 3 Pro:
- Cache Read: $0.26/M (90% cheaper than input)
- Use case: Repeated system prompts, documentation, knowledge bases
- Strategy: Place cacheable content at the start of your prompt
- 100K context without cache: $260/1M requests
- With cache: $26/1M requests = 90% savings
Context Windows
Gemini models support large context:- Gemini 3 Pro: Up to 2M tokens
- Gemini 2.5 Pro: Up to 2M tokens
- Gemini 2.5 Flash: Up to 1M tokens
Support
Need help with Gemini integration?Google AI Documentation
Official Gemini documentation
Google AI Pricing
Official pricing details
Splox Docs
Browse our guides
Community
Get help from the community

