WWhichAITry via Vercel AI Gateway
Models/Google
Google · released 2026-01

Gemini 3 Flash

Cheap, fast, 2M-token context.

context
2M
output max
32K
input /M
$0.30
output /M
$1.20
cached in /M
$0.07
modalities
T·I·A·V
Verdict

What it's actually good at

Strengths

  • +Unbeatable cost for the context window
  • +Very fast

Weaknesses

  • Reasoning behind Pro/Opus tier

Our scores

reasoning
7.0
coding
7.0
writing
7.0
speed
10.0
value
10.0
Flash tier with near-Pro quality on most workloads at a fraction of the cost. Excellent default for high-throughput multimodal apps.
Pricing

Cost at common volumes

Monthly volumeInputOutputEst. monthly
Side project1M tok0.3M tok$0.66
Growing app50M tok15M tok$33.00
Production500M tok150M tok$330.00
Scale5000M tok1500M tok$3300.00

Estimates assume uncached input. Prompt caching and batch APIs can cut this by 50–90% for many workloads. Use the calculator →

Head-to-head

Compare Gemini 3 Flash