Robots AtlasRobots Atlas
Gemini 2.5 Pro Logo
Active
Mar 25, 2025
APIHosted UICloud

Gemini 2.5 Pro

LLMMultimodalReasoningTool-Using

Gemini 2.5 Pro is Google DeepMind's flagship reasoning model, generally available June 17, 2025. Built on Sparse MoE architecture, supports up to 1M token context, text/audio/image/video input, and integrated thinking mode.

Technical specification

Context window
do 1M tokenów
Parameters
nieujawnione
Max output
0K
License
proprietary
Tools
Yes
Fine-tuning
No
Weights access
Closed
Knowledge cutoff
Jan 2025
Hardware requirementsAccess via Google Cloud infrastructure (Vertex AI / Gemini API)
Last updated: Apr 20, 2026

Modalities

Input
Text
Image
Audio
Video
Documents
Structured data
URLs
Output
Text
Code
Structured data
summaries
Analytical reports
Image

Capabilities

17

Reasoning

Reasoning

Multi-step reasoning

Reasoning

Long context

Reasoning

Coding

Coding

Function Calling

Planning

Structured output

Structured gen.

Audio understanding

Audio

Image understanding

Vision

Video Understanding

Other

Chart understanding

Vision

Diagram reasoning

Reasoning

OCR

Vision

Multilingual

Language

Planning

Planning

Streaming output

Reasoning

Interleaved Multimodal Input

Reasoning

Multimodal understanding

Multimodality

Applications

Pricing

PublicUSDper 1M tokens

Two-tier pricing based on context length. Prompts ≤200K tokens: $1.25/MTok input, $10.00/MTok output (thinking tokens counted as output). Prompts >200K tokens: $2.50/MTok input, $15.00/MTok output. Context caching: $0.31/MTok (≤200K), $0.625/MTok (>200K), storage $4.50/MTok/h. Batch API: ~50% discount. Free tier available in Google AI Studio (data used for product training).

Pricing · Calculator
Token volume10K

INPUT

$1.2500 / per 1M tokens

$0.0125

OUTPUT

$10.0000 / per 1M tokens

$0.1000

CACHE

$0.3100 / per 1M tokens

$0.00310

TOTAL

for 10K tokens

$0.1125
Price per 1M tokens · USD
Standard
All plans3
TextStandard/ per 1M tokens
Input$1.2500
Cache$0.3100
Output$10.0000

Output includes thinking tokens. Context caching: $0.31/MTok read, storage $4.50/MTok/h.

TextStandard/ per 1M tokens
Input$2.5000
Cache$0.6250
Output$15.0000

Higher rate applies for long contexts exceeding 200K tokens.

TextBatch/ per 1M tokens
Input$0.6250
Output$5.0000

Batch API (~50% discount). Asynchronous processing.

Ceny dotyczą płatnego tieru Gemini Developer API (Google AI Studio). Vertex AI ma osobny cennik (cloud.google.com/vertex-ai/generative-ai/pricing) — zasadniczo zbliżony. Thinking tokens liczone jako output tokens. Grounding with Google Search: 1500 darmowych zapytań/dzień, powyżej $35/1000 zapytań.