Skip to content
All Newsletters

FLI July 2022 Newsletter

August 1, 2022
Will Jones


Meet the First Recipients of FLI’s New Grants Program

This month we are delighted to announce the inaugural winners of our new AI safety research grants, funded by Vitalik Buterin. Out of a total of almost 100 applications, an external review panel of 25 leading academics selected these eight PhDs and one Postdoctoral Fellow. Let's meet them!


The Vitalik Buterin PhD Fellowship in AI Existential Safety is for students starting PhD programs in 2022 who plan to work on AI existential safety research, or for existing PhD students who would not otherwise have funding to work on AI existential safety research. It will fund students for 5 years of their PhD, with extension funding possible. At universities in the US, UK, or Canada, annual funding will cover tuition, fees, and the stipend of the student’s PhD program up to $40,000, as well as a fund of $10,000 that can be used for research-related expenses such as travel and computing. At universities not in the US, UK or Canada, the stipend amount will be adjusted to match local conditions. Fellows will also be invited to workshops where they will be able to interact with other researchers in the field.

These are the PhD Fellows this year, and their research focuses (pictures above, clockwise starting top left):

Usman Anwar - Research on inverse constrained reinforcement learning, impact regularizers, and distribution generalization, at the University of Cambridge.

Stephen Casper - Research on interpretable and robust AI at Massachusetts Institute of Technology (MIT).

Xin Cynthia Chen - Research on safe reinforcement learning at ETH Zurich.

Erik Jenner - Research on developing principled techniques for aligning even very powerful AI systems with human values at UC Berkeley.

Johannes Treutlein - Research on objective robustness and learned optimization at UC Berkeley.

Alexander Pan - Research on alignment, adversarial robustness, and anomaly detection at UC Berkeley.

Erik Jones - Research on robustness of forthcoming machine learning systems at UC Berkeley.

Zhijing Jin - Research on promoting NLP for social good and improving AI by connecting NLP with causality, at Max Planck Institute.

Postdoctoral Fellowship

The Vitalik Buterin Postdoctoral Fellowship in AI Existential Safety is designed to support promising researchers for postdoctoral appointments starting in the fall semester of 2022 to work on AI existential safety research. Funding is for three years subject to annual renewals based on satisfactory progress reports. For host institutions in the US, UK, or Canada, the Fellowship includes an annual $80,000 stipend and a fund of up to $10,000 that can be used for research-related expenses such as travel and computing. At universities not in the US, UK or Canada, the fellowship amount will be adjusted to match local conditions.

Nisan Stiennon is our first Postdoctoral Fellow for this program; Nisan is researching agent foundations at UC Berkeley's Center for Human-Compatible AI (CHAI).

Find out more about these amazing fellows and their research interests. We intend to run these fellowships again this coming fall. Any questions about the programs can be sent to


2022 Future of Life Award Announcement Coming Soon

On August 6th, FLI will announce the 2022 Winners of the Future of Life Award. Watch our new video to find out what this award is all about, and hear the stories of our previous winners. The award, funded by FLI co-founder Jaan Tallinn, is a $50,000/person prize given to individuals who, without having received much recognition at the time, have helped make today's world dramatically better than it may otherwise have been. Follow the award countdown so you don't miss the announcement of the winners next week!

FLI Event in New York on Hiroshima Day

Join us at Pioneer Works in Brooklyn, New York, on 6th August, the 77th anniversary of Hiroshima, to explore a chilling but fascinating topic: the latest science of nuclear winter, and how to reduce this risk by improving awareness of it. We’ll be celebrating the heroes who discovered and spread the word about this shocking scientific prediction of nuclear winter: that firestorms set off by a major nuclear war would envelop the earth in soot and smoke blocking sunlight for years, sending global temperature plunging, ruining ecosystems and agriculture and killing billions through famine.

One panel, featuring nuclear winter pioneers Alan RobockBrian Toon and Richard Turco and moderated by FLI's Emilia Javorsky, will discuss the most up-to-date findings about nuclear winter. The other, featuring nuclear winter pioneers John Birks and Georgiy Stenchikov as well as Ann Druyan, the Award-winning American documentary producer and director who co-wrote Cosmos with her late husband Carl Sagan, will explore the challenges involved in communicating this risk to politicians and the public, from the 1980s to today. The specter of nuclear war still hangs over us, and current geopolitics has made it more threatening than ever. So come along and hear about it from the experts. Get your tickets here.


Senior Job Opening on the FLI Policy Team

We have another new job opening at the Future of Life Institute, and it's a big one: We're looking for a new Director of U.S. Policy, based in Washington D.C., Sacramento, or San Francisco. FLI has recently grown to about 20 staff and we anticipate continued growth across our activities of outreach, grant making, and advocacy. The Director of U.S. Policy plays a pivotal role in the organisation, and we are looking for a driven individual of substantial experience to fill this position. If you think you're up to it, or know someone who is, take a look at the full information.

FLI AI Director Richard Mallah Co-Organises AI Safety Workshop

On July 24-25th, Richard Mallah, Director of AI Projects at FLI, co-organized the 'AISafety 2022' workshop at the 31st International Joint Conference on Artificial Intelligence and the 23rd European Conference on Artificial Intelligence (IJCAI-ECAI 2022), in Vienna. The workshop sought 'to explore new ideas on safety engineering, as well as broader strategic, ethical and policy aspects of safety-critical AI-based systems'. Speakers included scientist and 'Rebooting AI' author Gary Marcus, Elizabeth Adams, of the Stanford University Institute for Human Centered AI, USA, and Prof. Dr. Thomas A. Henzinger, of the Institute of Science and Technology, Austria.

FLI is a 501c(3) non-profit organisation, meaning donations are tax exempt in the United States.
If you need our organisation number (EIN) for your tax return, it's 47-1052538.
FLI is registered in the EU Transparency Register. Our ID number is 787064543128-10.

Our newsletter

Regular updates about the Future of Life Institute, in your inbox

Subscribe to our newsletter and join over 20,000+ people who believe in our mission to preserve the future of life.

Recent newsletters

Future of Life Institute Newsletter: 'Imagine A World' is out today!

Envisioning Positive Futures With AI, + Urgent Calls For AI Regulation Continue.
September 5, 2023

Future of Life Institute Newsletter: Hollywood Talks AI

Including our latest video on AI + nukes, and FLI cause areas on the big screen.
August 2, 2023

Future of Life Institute Newsletter: Our Most Realistic Nuclear War Simulation Yet

Welcome to the Future of Life Institute newsletter. Every month, we bring 28,000+ subscribers the latest news on how emerging […]
July 5, 2023
All Newsletters

Sign up for the Future of Life Institute newsletter

Join 40,000+ others receiving periodic updates on our work and cause areas.
cloudmagnifiercrossarrow-up linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram