Google has launched Gemini 3.1 Flash-Lite, its most cost-efficient model to date, targeting developers and enterprises that need scalable AI at minimal cost. Priced at $0.025 per million tokens and delivering 2.5x faster response times, Flash-Lite fills a clear gap between raw capability and practical affordability.

The Cheapest Frontier Model on the Market

At $0.025 per million input tokens, Gemini 3.1 Flash-Lite undercuts most comparable models from OpenAI, Anthropic, and Mistral. For applications making thousands of API calls daily โ€” customer service bots, content moderation, document summarization pipelines โ€” this pricing difference compounds quickly. A workload costing $500/month on GPT-4o-mini could run for under $50 on Flash-Lite.

2.5x Faster Response Times

Speed matters for user-facing applications. Flash-Lite achieves 2.5x faster response times compared to previous Flash variants, making it suitable for real-time use cases like chat interfaces, live transcription assistance, and interactive coding tools. Google attributes this to architectural optimizations and more efficient inference infrastructure on its TPU clusters.

Where It Fits in Google's Model Lineup

Flash-Lite sits at the bottom of Google's model hierarchy: ultra-low cost, high throughput, good-enough quality. Above it sit Flash (balanced), Pro (capable), and Ultra (frontier). Developers can now route requests to the appropriate tier based on task complexity โ€” reducing costs without sacrificing quality where it matters. For startups and scale-ups building AI-native products, Flash-Lite is worth serious evaluation as a backbone for high-volume, straightforward workloads.