Command Palette

Search for a command to run...

YepAPI
Google/v1/ai/google/gemini-2.5-flash-lite

Gemini 2.5 Flash Lite

Access Gemini 2.5 Flash Lite through one API key. Google's cheapest model.

Google's lightest Gemini model. Ultra-cheap inference for high-volume tasks with 1M context.

Context Window

1.0M tokens

Max Output

66K tokens

Input Price

$0.10 / 1M tokens

Output Price

$0.40 / 1M tokens

Try it live

Send a message and see Gemini 2.5 Flash Lite respond in real time.

POST /v1/ai/google/gemini-2.5-flash-liteLive testing
Loading...

Maximum tokens in the response.

Real-time token streaming

Hit "Run" to see the response

Strengths

Speed
Cost-effective
Large context
High volume

Quick start

Copy this snippet and start making calls with Gemini 2.5 Flash Lite.

const res = await fetch('https://api.yepapi.com/v1/ai/google/gemini-2.5-flash-lite', {
  method: 'POST',
  headers: {
    'x-api-key': 'YOUR_API_KEY',
    'Content-Type': 'application/json',
  },
  body: JSON.stringify({
    "messages": [
      {
        "role": "user",
        "content": "Explain API gateways in 2 sentences."
      }
    ],
    "maxTokens": 256
  }),
});
const data = await res.json();
console.log(data);

Why use Gemini 2.5 Flash Lite through YepAPI?

One API key for all models — no separate accounts
OpenAI SDK compatible — just change the base URL
No monthly minimums — pay per token
Switch models with one line of code
Full provider passthrough — citations, search results, and all extras included
Streaming and non-streaming support on every model
Works with Cursor, Claude, LangChain, and any LLM tool
Unified billing across all providers

Frequently asked questions

Google's lightest Gemini model. Ultra-cheap inference for high-volume tasks with 1M context.

Input tokens cost $0.10 per 1M tokens and output tokens cost $0.40 per 1M tokens through YepAPI. No monthly minimums — you only pay for what you use.

Sign up for a free API key, then send requests to the /v1/ai/google/gemini-2.5-flash-lite endpoint. YepAPI is OpenAI SDK compatible, so you can use it with any tool that supports OpenAI — just change the base URL.

Gemini 2.5 Flash Lite supports a 1.0M token context window with up to 66K output tokens per request.

Ready to use Gemini 2.5 Flash Lite?

Get your API key and start making calls in 30 seconds. No credit card required.

Explore more models