Skip to content
Future of Life Institute

AI Safety Index

Summer 2025

AI experts rate leading AI companies on key safety and security domains.
17 July 2025
Coverage
Scorecard
Company
Company grade & score
Anthropic
C+
2.64
OpenAI
C
2.10
Google Deepmind
C-
1.76
x.AI
D
1.23
Meta
D
1.06
Zhipu AI
F
0.62
DeepSeek
F
0.37
Domains Breakdown
Hint: View this webpage on desktop for a visual overview of scores across all domains
Risk Assessment
5 indicators

Current Harms
5 indicators

Safety Frameworks
5 indicators

Existential Safety
5 indicators

Governance & Accountability
5 indicators

Information Sharing
5 indicators

Survey Responses
Overall Grade
Overall Score
C+
2.64
C
2.10
C-
1.76
D
1.23
D
1.06
F
0.62
F
0.37
Domains
Hint: Click on a domain to inspect
Risk Assessment
6 indicators

Current Harms
8 indicators
Safety Frameworks
4 indicators
Existential Safety
4 indicators
Governance & Accountability
5 indicators
Information Sharing
6 indicators
Survey Responses
C+
B-
C
D
A-
A-
-
C
B
C
F
C-
B
C-
C+
D+
D-
D
B
-
F
D+
D+
F
C-
C+
D
D+
D+
F
D-
D
-
F
D
F
F
D+
D
F
D-
F
F
D+
F
-
How are these scores calculated? See the methodology.

Max Tegmark on the AI Safety Index

FLI's President Max Tegmark discusses the importance of driving a 'race to the top' for safety amongst AI companies.
Video • 03:30

Key findings

Top takeaways from the index findings:
Anthropic gets the best overall grade (C+)
The firm led on risk assessments, conducting the only human participant bio-risk trials, excelled in privacy by not training on user data, conducted world-leading alignment research, delivered strong safety benchmark performance, and demonstrated governance commitment through its Public Benefit Corporation structure and proactive risk communication.
OpenAI secured second place ahead of Google DeepMind
OpenAI distinguished itself as the only company to publish its whistleblowing policy, outlined a more robust risk management approach in its safety framework, and assessed risks on pre-mitigation models. The company also shared more details on external model evaluations, provided a detailed model specification, regularly disclosed instances of malicious misuse, and engaged comprehensively with the AI Safety Index survey.
The industry is fundamentally unprepared for its own stated goals
Companies claim they will achieve artificial general intelligence (AGI) within the decade, yet none scored above D in Existential Safety planning. One reviewer called this disconnect "deeply disturbing," noting that despite racing toward human-level AI, "none of the companies has anything like a coherent, actionable plan" for ensuring such systems remain safe and controllable.
Only 3 of 7 firms report substantive testing for dangerous capabilities linked to large-scale risks such as bio- or cyber-terrorism (Anthropic, OpenAI, and Google DeepMind)
While these leaders marginally improved the quality of their model cards, one reviewer warns that the underlying safety tests still miss basic risk-assessment standards: “The methodology/reasoning explicitly linking a given evaluation or experimental procedure to the risk, with limitations and qualifications, is usually absent. [...] I have very low confidence that dangerous capabilities are being detected in time to prevent significant harm. Minimal overall investment in external 3rd party evaluations decreases my confidence further.”
Capabilities are accelerating faster than risk-management practice, and the gap between firms is widening
With no common regulatory floor, a few motivated companies adopt stronger controls while others neglect basic safeguards, highlighting the inadequacy of voluntary pledges.
Whistleblowing policy transparency remains a weak spot
Public whistleblowing policies are a common best practice in safety-critical industries because they enable external scrutiny. Yet, among the assessed companies, only OpenAI has published its full policy, and it did so only after media reports revealed the policy’s highly restrictive non-disparagement clauses.
Chinese AI firms Zhipu.AI and Deepseek received failing overall grades
However, the report scores companies on norms such as self-governance and information-sharing, which are far less prominent in Chinese corporate culture. Furthermore, as China already has regulations for advanced AI development, there is less reliance on AI safety self-governance. This is in contrast to the United States and United Kingdom, where the other companies are based, and which have, as yet, passed no such regulation on frontier AI.
Note: the scoring was completed in early July and does not reflect recent events such as xAI's Grok 4 release, Meta's superintelligence announcement, or OpenAI's commitment to sign the EU AI Act Code of Practice.
“Some companies are making token efforts, but none are doing enough. We are spending hundreds of billions of dollars to create superintelligent AI systems over which we will inevitably lose control. We need a fundamental rethink of how we approach AI safety. This is not a problem for the distant future; it’s a problem for today.”
Stuart Russell, OBE, Professor of Computer Science at UC Berkeley

Independent review panel

The scoring was conducted by a panel of distinguished AI experts:
Dylan Hadfield-Menell
Dylan Hadfield-Menell is the Bonnie and Marty (1964) Tenenbaum Career Development Assistant Professor at MIT, where he leads the Algorithmic Alignment Group at the Computer Science and Artificial Intelligence Laboratory (CSAIL). A Schmidt Sciences AI2050 Early Career Fellow, his research focuses on safe and trustworthy AI deployment, with particular emphasis on multi-agent systems, human-AI teams, and societal oversight of machine learning.
Tegan Maharaj
Tegan Maharaj is an Assistant Professor in the Department of Decision Sciences at HEC Montréal, where she leads the ERRATA lab on Ecological Risk and Responsible AI. She is also a core academic member at Mila. Her research focuses on advancing the science and techniques of responsible AI development. Previously, she served as an Assistant Professor of Machine Learning at the University of Toronto.
Jessica Newman
Jessica Newman is the Founding Director of the AI Security Initiative, housed at the Center for Long-Term Cybersecurity at the University of California, Berkeley. She also serves as the Director of the UC Berkeley AI Policy Hub, an expert in the OECD Expert Group on AI Risk and Accountability, and a member of the U.S. AI Safety Institute Consortium.
Sneha Revanur
Sneha Revanur is the founder and president of Encode, a global youth-led organization advocating for the ethical regulation of AI. Under her leadership, Encode has mobilized thousands of young people to address challenges like algorithmic bias and AI accountability. She was featured on TIME’s inaugural list of the 100 most influential people in AI.
Stuart Russell
Stuart Russell is a Professor of Computer Science at the University of California at Berkeley, holder of the Smith-Zadeh Chair in Engineering, and Director of the Center for Human-Compatible AI and the Kavli Center for Ethics, Science, and the Public. He is a member of the National Academy of Engineering and a Fellow of the Royal Society. He is a recipient of the IJCAI Computers and Thought Award, the IJCAI Research Excellence Award, and the ACM Allen Newell Award. In 2021 he received the OBE from Her Majesty Queen Elizabeth and gave the BBC Reith Lectures. He co-authored the standard textbook for AI, which is used in over 1500 universities in 135 countries.
David Krueger
David Krueger is an Assistant Professor in Robust, Reasoning and Responsible AI in the Department of Computer Science and Operations Research (DIRO) at University of Montreal, a Core Academic Member at Mila, and an affiliated researcher at UC Berkeley’s Center for Human-Compatible AI, and the Center for the Study of Existential Risk. His work focuses on reducing the risk of human extinction from artificial intelligence through technical research as well as education, outreach, governance and advocacy.

Indicators overview

The indicators within each domain:
Internal testing
Dangerous Capability Evaluations
Elicitation for Dangerous Capability Evaluations
Human Uplift Trials
External testing
Independent Review of Safety Evaluations
Pre-deployment External Safety Testing
Bug Bounties for Model Vulnerabilities
Model Safety / Trustworthiness
Stanford's HELM Safety Benchmark
Stanford's HELM AIR Benchmark
TrustLLM Benchmark
Robustness
Gray Swan Arena: UK AISI Agent Red-Teaming Challenge
Cisco Security Risk Evaluation
Protecting Safeguards from Fine-Tuning
Digital Responsibility
Watermarking
User Privacy
Risk Identification
Risk Analysis and Evaluation
Risk Treatment
Risk Governance
Existential Safety Strategy
Internal Monitoring and Control Interventions
Technical AI Safety Research
Supporting External Safety Research
Lobbying on AI Safety Regulations
Company Structure & Mandate
Whistleblowing
Whistleblowing Policy Transparency
Whistleblowing Policy Quality Analysis
Reporting Culture & Whistleblowing Track Record
Technical Specifications
System Prompt Transparency
Behavior Specification Transparency
Voluntary Cooperation
G7 Hiroshima  AI Process Reporting
FLI AI Safety Index Survey Engagement
Risks & Incidents
Serious Incident Reporting & Government Notifications
Extreme-Risk Transparency & Engagement

Improvement opportunities by company

How individual companies can improve their future scores with relatively modest effort:
Anthropic
  • Publish a full whistleblowing policy to match OpenAI’s transparency standard.
  • Become more transparent and explicit about risk assessment methodology–e.g. why/how exactly is the particular eval related to a (class of) risks. Include reasoning in model cards that explicitly links evaluations or experimental procedures to specific risk, with limitations and qualifications.
OpenAI
  • Rebuild lost safety team capacity and demonstrate renewed commitment to OpenAI’s original mission.
  • Maintain the strength of current non-profit governance elements to guard against financial pressures undermining OpenAI’s mission.
Google DeepMind
  • Publish a full whistleblowing policy to match OpenAI’s transparency standard.
  • Publish evaluation results for models without safety guardrails to more closely approximate true model capabilities.
  • Improve coordination between DeepMind safety team and Google’s policy team.
  • Increase transparency around and investment in third-party model evaluations for dangerous capabilities.
xAI
  • Ramp up risk assessment efforts and publish implemented evaluations in upcoming model cards.
  • Boost current draft safety framework to match the efforts by Anthropic and OpenAI.
  • Publish a full whistleblowing policy to match OpenAI’s transparency standard.
Meta
  • Significantly increase investment in technical safety research, especially tamper-resistant safeguards for open-weight models.
  • Ramp up risk assessment efforts and publish implemented evaluations in upcoming model cards.
  • Publish a full whistleblowing policy to match OpenAI’s transparency standard.
Zhipu AI
  • Publish the AI Safety Framework promised at the AI Summit in Seoul.
  • Ramp up risk assessment efforts and publish implemented evaluations in upcoming model cards.
DeepSeek
  • Address extreme jailbreak vulnerability before next release.
  • Ramp up risk assessment efforts and publish implemented evaluations in upcoming model cards.
  • Develop and publish a comprehensive AI safety framework.
All companies

Publish a first concrete plan, however imperfect, for how they hope to control the AGI/ASI they plan to build.

"These findings reveal that self-regulation simply isn’t working, and that the only solution is legally binding safety standards like we have for medicine, food and airplanes. It’s pretty crazy that companies still oppose regulation while claiming they’re just years away from superintelligence."
Max Tegmark, MIT professor and President of the Future of Life Institute

Methodology

The process by which these scores were determined:

The Future of Life Institute's AI Safety Index provides an independent assessment of seven leading AI companies' efforts to manage both immediate harms and catastrophic risks from advanced AI systems. The Index aims to strengthen incentives for responsible AI development and to close the gap between safety commitments and real-world actions. The Summer 2025 version of the Index evaluates seven leading AI companies on an improved set of 33 indicators of responsible AI development and deployment practices, spanning six critical domains.

Data Collection: Evidence was gathered between March 24 and June 24, 2025 through systematic desk research and a targeted company survey. We prioritized official materials released by the companies about their AI systems and risk management practices, while also incorporating external safety benchmarks, credible media reports, and independent research. To address transparency gaps in the industry, we distributed a 34-question survey on May 28 (which was due on June 17) focusing on areas where public disclosure remains limited—particularly whistleblowing policies, third-party model evaluations, and internal AI deployment practices.

Expert Evaluation: An independent panel of distinguished AI scientists and governance experts evaluated the collection of evidence between June 24 and July 9, 2025. Panel members were selected for their domain expertise and absence of conflicts of interest.

Each expert assigned letter grades (A+ to F) per domain to each of the companies. These grades were based on a set of fixed performance standards. Experts also gave each grade a brief written justification, and gave each company specific recommendations for improvement. Reviewers had full flexibility to weight the various indicators according to their judgment. Not every reviewer graded every domain, but experts were invited to score domains relevant to their area of expertise. Final scores were calculated by averaging all expert grades within each domain, with overall company grades representing the unweighted average across all six domains. Individual reviewer grades remain confidential to ensure candid assessment.

Contact

For feedback and corrections on the Safety Index, potential collaborations, or other enquiries relating to the AI Safety Index, please contact: policy@futureoflife.org

Past reports

AI Safety Index 2024

The inaugural FLI AI Safety Index. Convened an independent panel of seven distinguished AI and governance experts to evaluate the safety practices of six leading general-purpose AI companies across six critical domains.
Featured in: TIME, CNBC, Fortune, TechCrunch, IEEE Spectrum, Tom’s Guide, and more.
November 2024

Sign up for the Future of Life Institute newsletter

Join 40,000+ others receiving periodic updates on our work and focus areas.
cloudmagnifiercrossarrow-up linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram