option
Home
News
Google’s latest AI model report lacks key safety details, experts say

Google’s latest AI model report lacks key safety details, experts say

April 28, 2025
58

On Thursday, weeks after launching its latest and most advanced AI model, Gemini 2.5 Pro, Google released a technical report detailing the results of its internal safety assessments. However, experts have criticized the report for its lack of detail, making it challenging to fully understand the potential risks associated with the model.

Technical reports are crucial in the AI world, offering insights—even if they're sometimes unflattering—that companies might not usually share publicly. These reports are generally viewed by the AI community as genuine efforts to foster independent research and enhance safety evaluations.

Google's approach to safety reporting differs from some of its competitors. The company only publishes technical reports once a model moves beyond the "experimental" phase. Moreover, Google omits certain "dangerous capability" evaluation results from these reports, saving them for a separate audit.

Despite this, several experts expressed disappointment with the Gemini 2.5 Pro report to TechCrunch, pointing out its scant coverage of Google's proposed Frontier Safety Framework (FSF). Google unveiled the FSF last year, aiming to pinpoint future AI capabilities that might lead to "severe harm."

"This report is very sparse, contains minimal information, and was released weeks after the model was already made public," Peter Wildeford, co-founder of the Institute for AI Policy and Strategy, told TechCrunch. "It's impossible to verify if Google is living up to its public commitments and thus impossible to assess the safety and security of their models."

Thomas Woodside, co-founder of the Secure AI Project, acknowledged the release of the report for Gemini 2.5 Pro but questioned Google's dedication to providing timely supplemental safety evaluations. He noted that Google last published dangerous capability test results in June 2024, for a model announced in February of that year.

Adding to the concerns, Google has not yet released a report for Gemini 2.5 Flash, a smaller, more efficient model announced last week. A spokesperson informed TechCrunch that a report for Flash is "coming soon."

"I hope this is a promise from Google to start publishing more frequent updates," Woodside told TechCrunch. "Those updates should include the results of evaluations for models that haven’t been publicly deployed yet, since those models could also pose serious risks."

While Google was among the first AI labs to propose standardized reports for models, it's not alone in facing criticism for a lack of transparency. Meta released a similarly brief safety evaluation for its new Llama 4 open models, and OpenAI chose not to publish any report for its GPT-4.1 series.

Google's assurances to regulators about maintaining high standards in AI safety testing and reporting add pressure to the situation. Two years ago, Google promised the U.S. government to publish safety reports for all "significant" public AI models "within scope," followed by similar commitments to other countries, pledging "public transparency" around AI products.

Kevin Bankston, a senior adviser on AI governance at the Center for Democracy and Technology, described the trend of sporadic and vague reports as a "race to the bottom" on AI safety.

"Combined with reports that competing labs like OpenAI have reduced their safety testing time before release from months to days, this meager documentation for Google’s top AI model tells a troubling story of a race to the bottom on AI safety and transparency as companies rush their models to market," he told TechCrunch.

Google has stated that, although not detailed in its technical reports, it conducts safety testing and "adversarial red teaming" for models before their release.

Updated 4/22 at 12:58 p.m. Pacific: Modified language around the technical report’s reference to Google’s FSF.

Related article
Imagen 4:谷歌最新AI圖像生成器 Imagen 4:谷歌最新AI圖像生成器 Google近日發表最新圖像生成AI模型「Imagen 4」,宣稱將為用戶帶來比前代Imagen 3更出色的視覺體驗。本週稍早在Google I/O 2025大會亮相的這款新模型,被譽為在畫質與多樣性方面取得重大突破。Google表示,Imagen 4特別擅長處理織物質感、水珠反光與動物毛髮等精細紋理,同時能輕鬆駕馭寫實與抽象風格。其輸出解析度最高可達2K,
谷歌Gemini代碼助手強化AI編程代理功能 谷歌Gemini代碼助手強化AI編程代理功能 Google旗下AI程式開發助手Gemini Code Assist近期推出全新「代理模式」功能,目前開放預覽體驗。在最新Cloud Next大會上,Google展示這些AI代理如何突破性處理複雜編程任務——從Google文件規格書直接生成完整應用程式,或輕鬆實現跨語言程式碼轉換。更令人驚豔的是,開發者現可在Android Studio等整合開發環境中直接啟
谷歌的人工智慧未來基金可能需要謹慎行事 谷歌的人工智慧未來基金可能需要謹慎行事 Google 的新 AI 投資計劃:監管審查下的戰略轉變Google 最近宣布設立 AI 未來基金(AI Futures Fund),這標誌著這家科技巨頭在其塑造人工智慧未來的征程中邁出了大膽的一步。該計劃旨在為初創公司提供急需的資金、早期接觸仍在開發中的尖端人工智慧模型,以及來自 Google 內部專家的指導。儘管這不是 Google 第一次涉足初創企業生
Comments (5)
0/200
AlbertWalker
AlbertWalker April 29, 2025 at 12:00:00 AM GMT

Google's report on Gemini 2.5 Pro is a bit of a letdown. I was expecting more juicy details about the safety assessments, but it feels like they're holding back. It's hard to trust the AI fully without knowing the full story. Maybe next time, Google? 🤔

CharlesThomas
CharlesThomas April 29, 2025 at 12:00:00 AM GMT

ジェミニ2.5プロのレポート、ちょっとがっかりですね。安全評価の詳細をもっと知りたかったのに、情報が少なすぎる。AIを完全に信頼するのは難しいです。次回はもっと詳しくお願いします!😅

BillyThomas
BillyThomas April 28, 2025 at 12:00:00 AM GMT

El informe de Google sobre Gemini 2.5 Pro es decepcionante. Esperaba más detalles sobre las evaluaciones de seguridad, pero parece que están ocultando información. Es difícil confiar en la IA sin conocer toda la historia. ¿Tal vez la próxima vez, Google? 🤔

JimmyGarcia
JimmyGarcia April 29, 2025 at 12:00:00 AM GMT

O relatório do Google sobre o Gemini 2.5 Pro é um pouco decepcionante. Esperava mais detalhes sobre as avaliações de segurança, mas parece que eles estão escondendo algo. É difícil confiar totalmente na IA sem saber toda a história. Talvez na próxima, Google? 🤔

WalterKing
WalterKing April 29, 2025 at 12:00:00 AM GMT

Der Bericht von Google über Gemini 2.5 Pro ist ein bisschen enttäuschend. Ich hatte mehr Details zu den Sicherheitsbewertungen erwartet, aber es scheint, als würden sie Informationen zurückhalten. Ohne die ganze Geschichte ist es schwer, der KI vollständig zu vertrauen. Vielleicht beim nächsten Mal, Google? 🤔

Back to Top
OR