Google Just Raised the AI Bar
Google launched Gemini 3.1 Ultra with a 2-million token context window – the largest of any public AI model right now.
What Is a 2M Token Context?
It means the model can process and reason over roughly 1.5 million words at once. Think entire codebases, legal documents, or full novels in one shot.
Native Multimodal From Day One
Gemini 3.1 Ultra handles text, image, audio, and video natively – no switching between models. One model, all formats.
Meet Gemini 3.1 Flash-Lite
Alongside Ultra, Google launched Flash-Lite: 2.5x faster responses, 45% faster output, at just $0.25 per million input tokens.
Who Should Use This?
Developers building AI agents, researchers processing large datasets, and enterprises needing deep document analysis will benefit most.
What This Means for OpenAI
This is Google’s strongest challenge to GPT-4o yet. The context window alone puts Gemini 3.1 Ultra in a category of its own.

GIPHY App Key not set. Please check settings