option
Home News Understanding Long Context Windows: Key Insights

Understanding Long Context Windows: Key Insights

release date release date April 10, 2025
views views 97

Yesterday, we unveiled our latest breakthrough in AI technology with the Gemini 1.5 model. This new iteration brings significant enhancements in speed and efficiency, but the real game-changer is its innovative long context window. This feature allows the model to process an unprecedented number of tokens — the fundamental units that make up words, images, or videos — all at once. To shed light on this advancement, we turned to the Google DeepMind project team for insights into what long context windows are and how they can revolutionize the way developers work.

Understanding long context windows is crucial because they enable AI models to maintain and recall information throughout a session. Imagine trying to remember a name just minutes after it's mentioned in a conversation, or rushing to write down a phone number before it slips your mind. AI models face similar challenges, often "forgetting" details after a few interactions. Long context windows address this issue by allowing the model to keep more information in its "memory."

Previously, the Gemini model could handle up to 32,000 tokens simultaneously. However, with the release of 1.5 Pro for early testing, we've pushed the boundaries to a staggering 1 million tokens — the largest context window of any large-scale foundation model to date. Our research has even gone beyond this, successfully testing up to 10 million tokens. The larger the context window, the more diverse and extensive the data — text, images, audio, code, or video — the model can process.

Nikolay Savinov, a Google DeepMind Research Scientist and one of the leads on the long context project, shared, "Our initial goal was to reach 128,000 tokens, but I thought aiming higher would be beneficial, so I proposed 1 million tokens. And now, our research has exceeded that by 10 times."

Achieving this leap required a series of deep learning innovations. Pranav Shyam's early explorations provided crucial insights that guided our research. Denis Teplyashin, a Google DeepMind Engineer, explained, "Each breakthrough led to another, opening up new possibilities. When these innovations combined, we were amazed at the results, scaling from 128,000 tokens to 512,000, then 1 million, and recently, 10 million tokens in our internal research."

The expanded capacity of 1.5 Pro opens up exciting new applications. For instance, instead of summarizing a document that's dozens of pages long, it can now handle documents thousands of pages in length. Where the previous model could analyze thousands of lines of code, 1.5 Pro can now process tens of thousands of lines at once.

Machel Reid, another Google DeepMind Research Scientist, shared some fascinating test results: "In one test, we fed the entire codebase into the model, and it generated comprehensive documentation for it, which was incredible. In another, it accurately answered questions about the 1924 film Sherlock Jr. after 'watching' the entire 45-minute movie."

1.5 Pro also excels at reasoning across data within a prompt. Machel highlighted an example involving the rare language Kalamang, spoken by fewer than 200 people worldwide. "The model can't translate into Kalamang on its own, but with the long context window, we could include the entire grammar manual and example sentences. The model then learned to translate from English to Kalamang at a level comparable to someone learning from the same material."

Gemini 1.5 Pro comes with a standard 128K-token context window, but a select group of developers and enterprise customers can access a 1 million token context window through AI Studio and Vertex AI in private preview. Managing such a large context window is computationally intensive, and we're actively working on optimizations to reduce latency as we scale it out.

Looking ahead, the team is focused on making the model faster and more efficient, with safety as a priority. They're also exploring ways to further expand the long context window, enhance underlying architectures, and leverage new hardware improvements. Nikolay noted, "10 million tokens at once is nearing the thermal limit of our Tensor Processing Units. We're not sure where the limit lies yet, and the model might be capable of even more as hardware continues to evolve."

The team is eager to see the innovative applications that developers and the broader community will create with these new capabilities. Machel reflected, "When I first saw we had a million tokens in context, I wondered, 'What do you even use this for?' But now, I believe people's imaginations will expand, leading to more creative uses of these new capabilities."

[ttpp][yyxx]

Related article
Google’s AI Futures Fund may have to tread carefully Google’s AI Futures Fund may have to tread carefully Google’s New AI Investment Initiative: A Strategic Shift Amid Regulatory ScrutinyGoogle's recent announcement of an AI Futures Fund marks a bold move in the tech giant's ongoing qu
Oura adds AI-powered glucose tracking and meal logging Oura adds AI-powered glucose tracking and meal logging Oura Reinforces Its Commitment to Metabolic Health with Two Exciting New FeaturesOura is stepping up its game in the world of metabolic health with two cutting-edge, AI-driven feat
Judge slams lawyers for ‘bogus AI-generated research’ Judge slams lawyers for ‘bogus AI-generated research’ Judge Penalizes Law Firms for Using AI Without DisclosureIn a recent ruling, California Judge Michael Wilner slapped two prominent law firms with a hefty fine of $31,000 for secret
Comments (25)
0/200
NicholasRoberts
NicholasRoberts April 14, 2025 at 10:59:46 PM GMT

Gemini 1.5's long context window is a game-changer, no doubt! But sometimes it feels like it's trying to do too much at once, which can slow things down. Still, for processing huge chunks of data, it's unbeatable. Worth checking out! 🚀

HenryJackson
HenryJackson April 13, 2025 at 5:36:58 PM GMT

Gemini 1.5の長いコンテキストウィンドウは本当に革新的!でも、時々一度にたくさんのことをしようとして遅くなることがあるよね。それでも、大量のデータを処理するには最強だと思う。試してみる価値あり!🚀

ScottJackson
ScottJackson April 12, 2025 at 4:31:28 PM GMT

Gemini 1.5의 긴 컨텍스트 윈도우는 정말 혁신적이에요! 하지만 때때로 너무 많은 일을 한꺼번에 하려다 느려지는 것 같아요. 그래도 대량의 데이터를 처리하는 데는 최고예요. 한번 써보세요! 🚀

MatthewGonzalez
MatthewGonzalez April 16, 2025 at 3:41:59 PM GMT

A janela de contexto longo do Gemini 1.5 é revolucionária, sem dúvida! Mas às vezes parece que está tentando fazer muito de uma vez, o que pode atrasar as coisas. Ainda assim, para processar grandes quantidades de dados, é imbatível. Vale a pena conferir! 🚀

StevenGreen
StevenGreen April 10, 2025 at 2:12:10 PM GMT

La ventana de contexto largo de Gemini 1.5 es un cambio de juego, sin duda. Pero a veces parece que intenta hacer demasiado a la vez, lo que puede ralentizar las cosas. Aún así, para procesar grandes cantidades de datos, es insuperable. ¡Vale la pena probarlo! 🚀

DouglasMartinez
DouglasMartinez April 14, 2025 at 4:35:33 PM GMT

Gemini 1.5's long context window is a game-changer! It's amazing how much more it can process compared to older models. Just wish it was a bit faster. Still, a big step forward! 💪

Back to Top
OR