Gemini Context Calculator

Calculate token usage against Gemini 1M and 2M context windows.

Max output for Gemini 1.5 Pro: 8,192 tokens
Total context usage
5000.0% of 2.0M
System tokens0
User tokens0
Output tokens500
Remaining1,999,500

Related Tools

FAQ

How large is Gemini 1.5 Pro context window?
Gemini 1.5 Pro supports a 2,000,000-token context window — the largest of any major commercial model. This is equivalent to roughly 1.5 million words, 3,000 pages of text, or an entire codebase.
What is the difference between Gemini 1.5 Pro and Flash context windows?
Gemini 1.5 Pro has a 2M token context window while Gemini 1.5 Flash supports 1M tokens. Both vastly exceed GPT-4o (128K) and Claude (200K). Flash is faster and cheaper but has half the context capacity.
Does Gemini charge more for larger context inputs?
Yes. Gemini 1.5 Pro charges $1.25 per million tokens for prompts up to 128K tokens, and $2.50 per million tokens for prompts above 128K. Gemini 1.5 Flash uses a similar tiered pricing model.

Check how your prompts fit within Gemini 1.5 Pro (2M tokens), Gemini 1.5 Flash (1M tokens), and Gemini 1.0 Pro (32K) context windows. Enter system and user prompts to see remaining capacity and detect overflows.