Skip to content

Policy work

We aim to improve AI governance over civilian applications, autonomous weapons and in nuclear launch.

Introduction

Shaping the long-term through near-term advocacy

The policy team at FLI works to reduce extreme, large-scale risks from transformative technologies by improving national and international governance of Artificial Intelligence (AI).

FLI has spearheaded numerous efforts to this end. Most notably, in 2017 we created the influential Asilomar AI principles, a set of governance principles signed by thousands of leading minds in AI research and industry. More recently, the UN Secretary General consulted FLI as the civil society ‘co-champion’ for AI recommendations on the Digital Cooperation Roadmap.

In the civilian domain, we advise the European Union on how to strengthen and future-proof their upcoming EU AI Act; and in the U.S., we focus on increasing federal spending on AI safety research and strengthening the NIST AI Risk Management Framework. In the military domain, we advocate for a treaty on autonomous weapons at the United Nations and inform policymakers about the risks of incorporating AI systems into nuclear launch.
Our work

Policy projects

Educating about Lethal Autonomous Weapons

Military AI applications are rapidly expanding. We develop educational materials about how certain narrow classes of AI-powered weapons can harm national security and destabilize civilization, notably weapons where kill decisions are fully delegated to algorithms.

Strengthening the European AI Act

Our key recommendations include broadening the Act’s scope to regulate general purpose systems and extending the definition of prohibited manipulation to include any type of manipulatory technique, and manipulation that causes societal harm.

Strengthening the NIST AI Risk Management Framework

Our feedback on the first draft of the National Institute of Standards and Technology’s (NIST) AI risk management framework addressed extreme and unacceptable risks, loyalty of AI systems and the risk management of general purpose systems.

Mitigating the Risks of AI Integration in Nuclear Launch

Avoiding nuclear war is in the national security interest of all nations. We pursue a range of initiatives to reduce this risk. Our current focus is on mitigating the emerging risk of AI integration into nuclear command, control and communication.
Our content

Latest policy papers

FLI Position Paper on AI Liability

November 2022

Emerging Non-European Monopolies in the Global AI Market

November 2022

Civil society letter GPAIS October 2022

October 2022

A Proposal for a Definition of General Purpose Artificial Intelligence Systems

October 2022

Geographical Focus

Where you can find us

We are a hybrid organisation. Most of our policy work takes place in the US (D.C. and California), the EU (Brussels) and at the UN (New York and Geneva).

United States

In the US, FLI works to increase federal spending on AI safety research and to strengthen the NIST AI Risk Management Framework.

European Union

In Europe, our focus is on strengthening the EU AI Act and encouraging European states to support a treaty on autonomous weapons.

United Nations

At the UN, FLI works to promote the adoption of a legally-binding instrument on autonomous weapons.
Key partners

Achievements

Some of the things we have achieved

Developed the AI Asilomar Principles

In 2017, FLI coordinated the development of the Asilomar AI Principles, one of the earliest and most influential sets of AI governance principles.
View the principles

AI recommendation in the UN digital cooperation roadmap

Our recommendations (3C) on the global governance of AI technologies were adopted in the UN Secretary-General's digital cooperation roadmap.
View the roadmap

Max Tegmark's testimony to the EU parliament

Our founder and board member Max Tegmark presented a testimony on the regulation of general-purpose AI systems in the EU parliament.
Watch the testimony
Our content

Featured posts

Here is a selection of posts relating to our policy work:

Global AI Policy

How countries and organizations around the world are approaching the benefits and risks of AI Artificial intelligence (AI) holds great […]
August 2, 2021

Contact us

Let's put you in touch with the right person.

We do our best to respond to all incoming queries within three business days. Our team is spread across the globe, so please be considerate and remember that the person you are contacting may not be in your timezone.

Sign up for the Future of Life Institute newsletter

Join 20,000+ others receiving periodic updates on our work and cause areas.
View previous editions
cloudmagnifiercrossarrow-up linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram