Back
Google Developers BlogUpdateGoogle Developers Blog2026-04-05

Google Adds Flex and Priority Tiers to Gemini API — Letting Developers Trade Latency for Cost

Alongside the Gemma 4 launch on April 2, Google introduced Flex and Priority inference tiers for the Gemini API. Flex tier is cheaper with variable latency — designed for batch workloads and async agents. Priority tier guarantees low latency for real-time applications. Developers can now explicitly declare which tradeoff they need rather than getting a one-size-fits-all API response.

Original source

Google quietly shipped one of the most developer-relevant API changes of the year alongside its Gemma 4 announcement on April 2, 2026: Flex and Priority inference tiers for the Gemini API. The change gives developers explicit control over the cost-latency-reliability triangle that previously required workarounds like request queuing and retry logic.

**Flex tier** is designed for workloads that don't need immediate responses — batch processing, async agent pipelines, background document analysis. It's priced lower than standard inference, with variable latency that reflects actual GPU availability rather than a guaranteed SLA. For developers running nightly batch jobs or processing large document sets, this can meaningfully reduce API spend.

**Priority tier** guarantees low latency for real-time applications — chat interfaces, voice assistants, live agent workflows. It's priced at a premium over standard but provides the reliability guarantees that production user-facing features need.

The practical impact is significant for agent developers. Most agentic pipelines include a mix of real-time steps (where users are waiting) and background steps (where they're not). Previously, developers paid Priority prices for everything or wrote custom orchestration to manage the mixed requirements. Flex/Priority tiers make the API do that work.

This is also a competitive response to the inference pricing wars. OpenAI, Anthropic, and Mistral have all adjusted pricing tiers in recent months. Google's approach of named tiers with explicit tradeoff semantics is cleaner than pure price-per-token for developer decision-making.

Panel Takes

The Builder

The Builder

Developer Perspective

This is the API design I've wanted for two years. Running async pipelines on Priority tier pricing is wasteful — Flex tier at lower cost for background work and Priority for user-facing steps lets you build economically sound agents without custom queuing logic. This should be the standard across all inference providers.

The Skeptic

The Skeptic

Reality Check

Flex tier pricing and latency guarantees aren't fully disclosed yet, and 'variable latency' in production could mean anything from 2x slower to 10x slower during peak demand. Until independent developers publish real-world Flex tier latency distributions, it's hard to know if this is genuinely useful or just a price tier rename.

The Futurist

The Futurist

Big Picture

The commoditization of inference is accelerating — and when inference is a commodity, the differentiator becomes developer experience and API ergonomics. Flex/Priority tiers are Google acknowledging that agentic workloads have fundamentally different requirements than one-shot queries. This is infrastructure maturing.