option
Home
News
2024: AI's Year of Remarkable Growth and Innovation

2024: AI's Year of Remarkable Growth and Innovation

April 10, 2025
79

2024: AI

As we step into 2025, it's thrilling to reflect on the incredible strides we made in 2024. From launching the Gemini 2.0 models tailored for the agentic era to empowering creative expression, and from designing novel protein binders to advancing AI-enabled neuroscience and quantum computing, we've been pushing the boundaries of artificial intelligence responsibly and boldly. All of these efforts are aimed at harnessing AI for the greater good of humanity.

As we mentioned in our essay *Why we focus on AI* two years ago, our approach to AI development is rooted in our founding mission to organize the world’s information and make it universally accessible and useful. This mission drives our commitment to enhance the lives of as many people as possible, a goal that remains our north star.

In our 2024 Year-in-Review, we celebrate the remarkable achievements of the many talented teams at Google who have worked tirelessly to advance our mission. Their efforts have set the stage for even more exciting developments in the year ahead.

Relentless Innovation in Models, Products, and Technologies

2024 was all about experimentation, rapid deployment, and getting our latest technologies into the hands of developers. In December, we unveiled the first models of our Gemini 2.0 experimental series, designed specifically for the agentic era. We kicked things off with Gemini 2.0 Flash, our versatile workhorse, followed by cutting-edge prototypes from our agentic research. These include an updated Project Astra, exploring the potential of a universal AI assistant; Project Mariner, an early prototype capable of performing actions in Chrome as an experimental extension; and Jules, an AI-powered code agent. We're eager to integrate Gemini 2.0's capabilities into our flagship products, and we've already begun testing in AI Overviews within Search, used by over a billion people to explore new types of questions.

We also rolled out Deep Research, a new agentic feature in Gemini Advanced that saves hours of research by creating and executing multi-step plans for answering complex questions. Additionally, we introduced Gemini 2.0 Flash Thinking Experimental, a model that transparently displays its thought process.

Earlier in the year, we made significant strides by integrating Gemini's capabilities into more Google products and launching Gemini 1.5 Pro and Gemini 1.5 Flash. The latter, optimized for speed and efficiency, became our most popular model among developers thanks to its compact size and cost-effectiveness.

We also enhanced AI Studio, making it available as a progressive web app (PWA) installable on desktop, iOS, and Android, providing developers with a robust set of resources. The public's response to new features in NotebookLM, such as Audio Overviews, has been fantastic. These features can generate deep dive discussions from uploaded source material, making learning more engaging.

Speech input and output continue to be refined in products like Gemini Live, Project Astra, Journey Voices, and YouTube's auto dubbing, enhancing user interaction.

In line with our tradition of contributing to the open community, we released two new models from Gemma, our state-of-the-art open model, built on the same research and technology as Gemini. Gemma outperformed similarly sized models in areas like question answering, reasoning, and coding. We also released Gemma Scope, a tool to help researchers understand Gemma 2's inner workings.

We made strides in improving the factuality of our models and reducing hallucinations. In December, we published FACTS Grounding, a benchmark developed in collaboration with Google DeepMind, Google Research, and Kaggle, to evaluate how well large language models ground their responses in provided source material and avoid hallucinations.

The FACTS Grounding dataset, with 1,719 examples, is designed to test long-form responses grounded in context documents. We tested leading LLMs using FACTS Grounding, and we're proud to report that Gemini 2.0 Flash Experimental, Gemini 1.5 Flash, and Gemini 1.5 Pro achieved the top three factuality scores, with Gemini-2.0-flash-exp scoring an impressive 83.6%.

We also improved ML efficiency through innovative techniques like blockwise parallel decoding, confidence-based deferral, and speculative decoding, which speed up the inference times of LLMs. These improvements benefit Google products and set industry standards.

In sports, we launched TacticAI, an AI system for football tactics that provides tactical insights, particularly on corner kicks.

Our commitment to research leadership remains strong. A 2010-2023 WIPO survey on Generative AI citations showed that Google, including Google Research and Google DeepMind, received more than double the citations of the second-most cited institution.

This WIPO graph, based on January 2024 data from The Lens, highlights Alphabet's significant contributions to generative AI research over the past decade. Finally, we made progress with Project Starline, our "magic window" technology, partnering with HP to commercialize it, aiming to integrate it into video conferencing services like Google Meet and Zoom.

Empowering Creative Vision with Generative AI

We believe AI can unlock new realms of creativity, making creative expression more accessible and helping people realize their artistic visions. In 2024, we introduced a series of updates to our generative media tools, covering images, music, and video.

At the start of the year, we launched ImageFX and MusicFX, generative AI tools that create images and up-to-70-second audio clips from text prompts. At I/O, we previewed MusicFX DJ, designed to make live music creation more accessible. In October, we worked with Jacob Collier to simplify MusicFX DJ for new and aspiring musicians. We also updated our music AI toolkit, Music AI Sandbox, and evolved our Dream Track experiment, allowing U.S. creators to generate instrumental soundtracks across various genres using text-to-music models.

Later in the year, we released Veo 2 and Imagen 3, our latest image and video models. Imagen 3, our highest quality text-to-image model, generates images with superior detail, lighting, and fewer artifacts. Veo 2 demonstrated a better understanding of real-world physics and human movement, enhancing realism. Veo 2 marks a significant advancement in high-quality video generation.

We continued to explore AI's potential in editing, using it to control attributes like transparency and roughness of objects.

These examples showcase AI's ability to edit material properties using synthetic data generation.

In audio generation, we improved video-to-audio (V2A) technology, generating dynamic soundscapes from text prompts based on on-screen action, which can be paired with AI-generated video from Veo.

Games offer a perfect playground for creative exploration and training embodied agents. In 2024, we introduced Genie 2, a foundation world model that generates diverse, playable 3D environments for training and evaluating embodied agents. This followed the launch of SIMA, which can follow natural-language instructions in various video game settings.

The Architecture of Intelligence: Advances in Robotics, Hardware, and Computing

As our multimodal models become more adept at understanding the world's physics, they're enabling exciting advances in robotics. We're getting closer to our goal of more capable and helpful robots.

With ALOHA Unleashed, our robot mastered tasks like tying shoelaces, hanging shirts, repairing other robots, inserting gears, and cleaning kitchens.

At the year's start, we introduced AutoRT, SARA-RT, and RT-Trajectory, extensions of our Robotics Transformers work to help robots better navigate their environments and make quicker decisions. We also released ALOHA Unleashed, teaching robots to coordinate two arms, and DemoStart, which uses reinforcement learning to improve real-world performance on a multi-fingered robotic hand using simulations.

Robotic Transformer 2 (RT-2) learns from both web and robotics data, enabling it to perform tasks like placing a strawberry in a bowl. Beyond robotics, our AlphaChip reinforcement learning method is revolutionizing chip floorplanning for data centers and smartphones. We released a pre-trained checkpoint to facilitate external adoption of AlphaChip's open-source release. We also made Trillium, our sixth-generation TPU, available to Google Cloud customers, showcasing how AI can enhance chip design. AlphaChip learns to optimize chip layouts, improving with each design it creates.

Our research also tackled error correction in quantum computers. In November, we launched AlphaQubit, an AI-based decoder that identifies quantum computing errors with high accuracy. This collaboration between Google DeepMind and Google Research accelerated progress toward reliable quantum computers. In tests, AlphaQubit reduced errors by 6% compared to tensor network methods and by 30% compared to correlated matching.

In December, the Google Quantum AI team unveiled Willow, our latest quantum chip. Willow can perform a benchmark computation in under five minutes that would take today's fastest supercomputers 10 septillion years. Using quantum error correction, Willow halved the error rate, achieving a milestone known as "below threshold" and earning the Physics Breakthrough of the Year award.

Willow showcases state-of-the-art performance in quantum computing.

Uncovering New Solutions: Progress in Science, Biology, and Mathematics

We continued to accelerate scientific progress with AI, releasing tools and papers that demonstrate AI's power in advancing science and mathematics. Here are some highlights:

In January, we introduced AlphaGeometry, an AI system for solving complex geometry problems. Our updated AlphaGeometry 2 and AlphaProof, a reinforcement-learning-based system for formal math reasoning, achieved silver medalist performance at the July 2024 International Mathematical Olympiad.

AlphaGeometry 2 solved Problem 4 of the July 2024 International Mathematical Olympiad in just 19 seconds, proving that ∠KIL + ∠XPY equals 180°. In collaboration with Isomorphic Labs, we introduced AlphaFold 3, which predicts the structure and interactions of life's molecules, aiming to transform our understanding of biology and drug discovery.

AlphaFold 3's advanced architecture and training cover all life's molecules, from proteins to DNA. We also made significant strides in protein design with AlphaProteo, an AI system for creating high-strength protein binders, which could lead to new drugs and biosensors.

AlphaProteo can design new protein binders for various target proteins.

In collaboration with Harvard's Lichtman Lab, we produced a nano-scale mapping of a piece of the human brain, a first of its kind, and made it available for researchers. This follows our decade-long effort in connectomics, now extending to human brain mapping.

This brain mapping project reveals mirror-image cell clusters in the deepest layer of the cortex.

In late November, we co-hosted the AI for Science Forum with the Royal Society, discussing key topics like protein structure prediction, human brain mapping, and using AI for forecasting and wildfire detection. We also hosted a Q&A with four Nobel Laureates at the forum, available on the Google DeepMind podcast.

2024 was also a landmark year as Demis Hassabis, John Jumper, and David Baker received the Nobel Prize in Chemistry for their work on AlphaFold 2, recognized for revolutionizing protein design. Geoffrey Hinton, along with John Hopfield, received the Nobel Prize in Physics for foundational work in machine learning with artificial neural networks.

Google also received additional accolades, including the NeurIPS 2024 Test of Time Paper Awards and the Beale—Orchard-Hays Prize for Primal-Dual Linear Programming (PDLP), now part of Google OR Tools, aiding in large-scale linear programming with real-world applications.

AI for the Benefit of Humanity

This year, we made significant product advances and published research demonstrating how AI can directly and immediately benefit people in areas like healthcare, disaster readiness, and education.

In healthcare, AI promises to democratize quality care, particularly in early detection of cardiovascular disease. Our research showed that a simple fingertip device, combined with basic metadata, can predict heart health risks. We also advanced AI-enabled diagnostics for tuberculosis, showing how AI can screen populations with high TB and HIV rates effectively.

Med-Gemini achieved a new state-of-the-art score on the MedQA benchmark, surpassing our previous best, Med-PaLM 2, by 4.6%. Our Gemini model is a versatile tool for professionals, and we're developing fine-tuned models for specific domains. Med-Gemini, for instance, combines training on de-identified medical data with Gemini's capabilities, achieving 91.1% accuracy on the MedQA USMLE-style question benchmark.

We're also exploring how machine learning can address shortages in imaging expertise in fields like radiology, dermatology, and pathology. We released Derm Foundation and Path Foundation for diagnostic tasks and biomarker discovery, collaborated with Stanford Medicine on the Skin Condition Image Network (SCIN) dataset, and unveiled CT Foundation for medical imaging research.

In education, we introduced LearnLM, a family of models fine-tuned for learning, enhancing experiences in Search, YouTube, and Gemini. LearnLM outperformed other leading AI models, and we made it available to developers in AI Studio. Our conversational learning companion, LearnAbout, and the audio discussion tool, Illuminate, further enrich learning experiences.

In disaster forecasting and preparedness, we introduced GenCast, improving weather and extreme event forecasting, and NeuralGCM, capable of simulating thousands of days of atmospheric conditions. GraphCast, which won the 2024 MacRobert Award, provides detailed weather predictions.

GraphCast's predictions over 10 days showcase specific humidity, surface temperature, and wind speed.

We improved our flood forecasting model to predict flooding seven days in advance, expanding coverage to 100 countries and 700 million people.

Our flood forecasting model now covers over 100 countries, with virtual gauges in 150 countries where physical gauges are unavailable. AI also aids in wildfire detection and mitigation. Our Wildfire Boundary Maps are now available in 22 countries, and we created FireSat, a satellite constellation that can detect small wildfires within 20 minutes.

We expanded Google Translate to include 110 new languages, helping break down barriers to information and opportunity for over 614 million speakers.

These new languages in Google Translate represent 8% of the world's population.

Helping Set the Standard in Responsible AI

We continued our industry-leading research in AI safety, developing new tools and techniques and integrating these into our latest models. We're committed to collaborating to address risks.

Our research into misuse found that deep fakes and jailbreaks are the most common issues. In May, we introduced The Frontier Safety Framework to identify emerging capabilities in our advanced AI models and launched our AI Responsibility Lifecycle framework. In October, we expanded our Responsible GenAI Toolkit to work with any LLM, helping developers build AI responsibly.

We released a paper on The Ethics of Advanced AI Assistants, examining the technical and moral landscape of AI assistants and the opportunities and risks they pose.

We expanded SynthID's capabilities to watermark AI-generated text in the Gemini app and web experience and video in Veo. To enhance online transparency, we joined the Coalition for Content Provenance and Authenticity (C2PA) and worked on a new, more secure version of the Content Credentials standard.

SynthID adjusts the probability scores of predicted tokens to ensure quality, accuracy, and creativity in AI-generated content.

Beyond LLMs, we shared our biosecurity approach for AlphaFold 3, worked with industry partners to launch the Coalition for Secure AI (CoSAI), and participated in the AI Seoul Summit to contribute to international AI governance.

As we develop new technologies like AI agents, we'll continue to explore safety, security, and privacy questions. Guided by our AI Principles, we're taking a deliberate, gradual approach, conducting extensive research, safety training, and risk assessments with trusted testers and external experts.

Looking Ahead to 2025

2024 was a year of incredible progress and excitement in AI. We're even more thrilled about what's coming in 2025.

As we continue to push the boundaries of AI research in products, science, health, and creativity, we must thoughtfully consider how and when to deploy these technologies. By prioritizing responsible AI practices and fostering collaboration, we'll continue to play a crucial role in building a future where AI benefits humanity.

Related article
New Study Reveals How Much Data LLMs Actually Memorize New Study Reveals How Much Data LLMs Actually Memorize How Much Do AI Models Actually Memorize? New Research Reveals Surprising InsightsWe all know that large language models (LLMs) like ChatGPT, Claude, and Gemini are trained on enormous datasets—trillions of words from books, websites, code, and even multimedia like images and audio. But what exactly
AI’s potential to tackle the UK’s productivity puzzle AI’s potential to tackle the UK’s productivity puzzle AI Could Boost UK Economy by £400bn – But Only If Workers Embrace ItThe UK is on the brink of a massive economic opportunity—£400 billion in growth powered by AI. But here’s the catch: half of that potential depends on workers actually adopting AI tools. While interest in AI has skyrocketed, many pe
9 Best AI Recruiting Tools (April 2025) 9 Best AI Recruiting Tools (April 2025) The 10 Best AI Recruiting Tools Transforming Hiring in 2024Artificial intelligence is shaking up the recruitment game—and for good reason. Companies drowning in resumes and endless interview scheduling are turning to AI-powered solutions that can slash hiring time, eliminate bias, and surface the be
Comments (25)
0/200
JustinMartin
JustinMartin April 12, 2025 at 12:00:00 AM GMT

2024 was a wild ride with AI! From Gemini 2.0 to quantum computing, it's like we're living in a sci-fi movie. Loved seeing all the creative stuff AI can do, but sometimes it felt a bit too much, you know? Can't wait to see what 2025 brings, but maybe slow down a bit, AI?

BillyEvans
BillyEvans April 12, 2025 at 12:00:00 AM GMT

¡2024 fue increíble con la IA! Desde Gemini 2.0 hasta la computación cuántica, parece que vivimos en una película de ciencia ficción. Me encantó ver todo lo creativo que puede hacer la IA, pero a veces se sintió un poco abrumador, ¿sabes? Espero con ansias lo que 2025 traiga, pero tal vez, ¿puede la IA ralentizar un poco?

WalterWhite
WalterWhite April 10, 2025 at 12:00:00 AM GMT

2024年はAIがすごかったね!ジェミニ2.0から量子コンピューティングまで、まるでSF映画に住んでいるみたい。AIができるクリエイティブなことが大好きだけど、時々少し多すぎる感じがするんだよね。2025年が楽しみだけど、AI、ちょっとペースを落とそうか?

WillieRodriguez
WillieRodriguez April 11, 2025 at 12:00:00 AM GMT

2024 war ein wilder Ritt mit KI! Von Gemini 2.0 bis zur Quantencomputing, es fühlt sich an, als lebten wir in einem Sci-Fi-Film. Ich liebe das kreative Zeug, was KI kann, aber manchmal war es ein bisschen zu viel, weißt du? Ich freue mich auf 2025, aber vielleicht sollte die KI ein bisschen langsamer machen?

KeithJones
KeithJones April 12, 2025 at 12:00:00 AM GMT

Năm 2024 thật tuyệt vời với trí tuệ nhân tạo! Từ Gemini 2.0 đến máy tính lượng tử, cảm giác như chúng ta đang sống trong phim khoa học viễn tưởng. Mình thích những điều sáng tạo mà AI có thể làm, nhưng đôi khi nó hơi quá sức, bạn biết đấy? Mong chờ năm 2025, nhưng có lẽ AI nên chậm lại một chút?

StephenRamirez
StephenRamirez April 10, 2025 at 12:00:00 AM GMT

2024 was wild! AI's growth was insane, from Gemini 2.0 to protein binders. It's like every week there was something new and mind-blowing. Can't wait to see what 2025 brings, but 2024 set the bar high!

Back to Top
OR