Gemini 2.5 Flash Lite
Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, "thinking" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the [Reasoning API parameter](https://openrouter.ai/docs/use-cases/reasoning-tokens) to selectively trade off cost for intelligence.
The date this AI finished learning. It may not know about things that happened after this date.
The types of content this AI can receive, and what it can produce in return.
The maximum amount of text the AI can read and process in a single request. A larger number means it can handle longer documents or conversations.
The cost of using this AI directly in your own application. Shown in USD per 1 million units of text (tokens).