Skip to content

AI Safety Index Released

The Future of Life Institute has released its first safety scorecard of leading AI companies, finding many are not addressing safety concerns while some have taken small initial steps in the right direction.
Published:
December 11, 2024
Author:
Chase Hardin
Image: Cover page of the FLI AI Safety Index 2024.

For Immediate Release
December 11th, 2024
Media Contact: Chase Hardin, chase@futureoflife.org
+1 (623)986-0161


Major AI Companies Have ‘Significant Gaps’ in Safety Measures Say Leading AI Experts in External Safety Review

CAMPBELL, CA  – Today, the Future of Life Institute (FLI) released its 2024 AI Safety Index, in which several of the world’s foremost AI and governance experts assessed the safety standards of six prominent companies developing AI, including Anthropic, Google DeepMind, Meta, OpenAI, x.AI, and Zhipu AI. The independent panel evaluated each company in six categories: Risk Assessment, Current Harms, Safety Frameworks, Existential Safety Strategy, Governance & Accountability, and Transparency & Communication.

The review panel found that, while some companies demonstrated commendable practices in select domains, there are significant risk management disparities between them. All of the flagship models were found to be vulnerable to adversarial attacks, and despite their explicit ambitions to develop systems that rival or exceed human intelligence, firms have no adequate strategy for ensuring such systems remain beneficial and under human control.

“It’s horrifying that the very companies whose leaders predict AI could end humanity have no strategy to avert such a fate,” said panelist David Krueger, Assistant Professor at Université de Montreal and a core member of Mila.

“The findings of the AI Safety Index project suggest that although there is a lot of activity at AI companies that goes under the heading of ‘safety,’ it is not yet very effective,” said panelist Stuart Russell, a Professor of Computer Science at UC Berkeley. “In particular, none of the current activity provides any kind of quantitative guarantee of safety; nor does it seem possible to provide such guarantees given the current approach to AI via giant black boxes trained on unimaginably vast quantities of data. And it’s only going to get harder as these AI systems get bigger. In other words, it’s possible that the current technology direction can never support the necessary safety guarantees, in which case it’s really a dead end.”

Grading: Uses the US GPA system for grade boundaries: A+, A, A-, B+, […], F letter values corresponding to numerical values 4.3, 4.0, 3.7, 3.3, […], 0.

The final report can be viewed here.

“Evaluation initiatives like this Index are very important because they can provide valuable insights into the safety practices of leading AI companies. They are an essential step in holding firms accountable for their safety commitments and can help highlight emerging best practices and encourage competitors to adopt more responsible approaches,” said Professor Yoshua Bengio, Full Professor at Université de Montréal, Founder and Scientific Director of Mila – Quebec AI Institute and 2018 A.M. Turing Award co-winner.

Grades were assessed based on publicly available information as well as the companies’ responses to a survey conducted by FLI.  The review raised concerns that ongoing competitive pressures are encouraging companies to ignore or sidestep questions around the risks posed by developing this technology, resulting in significant gaps in safety measures and a serious need for improved accountability.

“We launched the Safety Index to give the public a clear picture of where these AI labs stand on safety issues,” said FLI president Max Tegmark, a professor doing AI research at MIT. “The reviewers have decades of combined experience in AI and risk assessment, so when they speak up about AI safety, we should pay close attention to what they say.”

Review panelists: 

  • Yoshua Bengio, Professor at Université de Montreal and Founder of Mila – Quebec Artificial Intelligence Institute. He is the recipient of the 2018 A.M. Turing Award.
  • Atoosa Kasirzadeh, an Assistant Professor at Carnegie Mellon University and a 2024 Schmidt Sciences AI2050 Fellow.
  • David Krueger, Assistant Professor at Université de Montreal and a core member of Mila and the Center for Human-compatible AI.
  • Tegan Maharaj, Assistant Professor at HEC Montréal and core faculty of Mila. She leads the ERRATA lab on Responsible AI.
  • Jessica Newman, Director of the AI Security Initiative at UC Berkeley and the Co-Director of the UC Berkeley AI Policy Hub.
  • Sneha Revanur, founder of youth AI advocacy organization Encode Justice and a Forbes 30 Under 30 honoree.
  • Stuart Russell is a Professor of Computer Science at UC Berkeley, where he leads the Center for Human-compatible AI. He co-authored the standard AI textbook used in more than 1,500 universities in 135 countries.

The Future of Life Institute is a global non-profit organization working to steer the development of transformative technologies towards benefiting life and away from extreme large-scale risks. To find out more about our mission or explore our work, visit www.futureoflife.org.

This content was first published at futureoflife.org on December 11, 2024.

About the Future of Life Institute

The Future of Life Institute (FLI) is a global non-profit with a team of 20+ full-time staff operating across the US and Europe. FLI has been working to steer the development of transformative technologies towards benefitting life and away from extreme large-scale risks since its founding in 2014. Find out more about our mission or explore our work.

Our content

Related content

Other posts about , ,

If you enjoyed this content, you also might also be interested in:

Max Tegmark on AGI Manhattan Project

A new report for Congress recommends that the US start a "Manhattan Project" to build Artificial General Intelligence. To do so would be a suicide race.
20 November, 2024

FLI Statement on White House National Security Memorandum

Last week the White House released a National Security Memorandum concerning AI governance and risk management. The NSM issues guidance […]
28 October, 2024

Paris AI Safety Breakfast #3: Yoshua Bengio

The third of our 'AI Safety Breakfasts' event series, featuring Yoshua Bengio on the evolution of AI capabilities, loss-of-control scenarios, and proactive vs reactive defense.
16 October, 2024

Paris AI Safety Breakfast #2: Dr. Charlotte Stix

The second of our 'AI Safety Breakfasts' event series, featuring Dr. Charlotte Stix on model evaluations, deceptive AI behaviour, and the AI Safety and Action Summits.
14 October, 2024
Our content

Sign up for the Future of Life Institute newsletter

Join 40,000+ others receiving periodic updates on our work and cause areas.
cloudmagnifiercrossarrow-up linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram