FLI September, 2019 Newsletter

NOT COOL: A Climate Podcast, Preparing for Global Catastrophe & More

In case you missed it: We’ve launched a new podcast…

Climate change, to state the obvious, is a huge and complicated problem. That’s why we’ve launched a new series in our podcast line-up: Not Cool: A Climate Podcast. We started this series because the news about climate change seems to get worse with each new article and report — but the solutions, at least as reported, remain vague and elusive. In this new series, hosted by Ariel Conn, we’ll hear directly from climate experts from around the world as they answer every question we can think of about the climate crisis.

You can find a short trailer here that highlights what we’ll be covering in the coming months. And of course you can jump right in to the first episode — we’ve listed the available episodes below, and they can also be found at futureoflife.org/notcool. You can also always listen to all FLI podcasts on any of your favorite podcast platforms just by searching for “Future of Life Institute.” The Not Cool podcasts are all there, and we’ll be releasing new episodes every Tuesday and Thursday for at least the next couple of months. We hope these interviews will help you better understand the science and policies behind the climate crisis and what we can all do to prevent the worst effects of climate change.

This week, we’ll be focusing on extreme events and compounds events. Tuesday’s episode will feature Stephanie Herring of the National Oceanic & Atmospheric Administration; On Thursday, we’ll be joined Jakob Zscheischler from Switzerland’s University of Bern. Future topics include ocean acidification, climate economics, adverse health effects of climate change, how to adapt, using machine learning to tackle climate change, and much more.

More Podcast Episodes

feeding everyone, global catastrophe

FLI Podcast: Feeding Everyone in a Global Catastrophe with Dave Denkenberger & Joshua Pearce

Most of us working on catastrophic and existential threats focus on trying to prevent them — not on figuring out how to survive the aftermath. But what if, despite everyone’s best efforts, humanity does undergo such a catastrophe? This month’s podcast is all about what we can do in the present to ensure humanity’s survival in a future worst-case scenario. Ariel is joined by Dave Denkenberger and Joshua Pearce, co-authors of the book Feeding Everyone No Matter What, who explain what would constitute a catastrophic event, what it would take to feed the global population, and how their research could help address world hunger today. They also discuss infrastructural preparations, appropriate technology, and why it’s worth investing in these efforts. Listen here.

AI Alignment Podcast: Synthesizing a human’s preferences into a utility function with Stuart Armstrong

In his Research Agenda v0.9: Synthesizing a human’s preferences into a utility function, Stuart Armstrong develops an approach for generating friendly artificial intelligence. His alignment proposal can broadly be understood as a kind of inverse reinforcement learning where most of the task of inferring human preferences is left to the AI itself. It’s up to us to build the correct assumptions, definitions, preference learning methodology, and synthesis process into the AI system such that it will be able to meaningfully learn human preferences and synthesize them into an adequate utility function. In order to get this all right, his agenda looks at how to understand and identify human partial preferences, how to ultimately synthesize these learned preferences into an “adequate” utility function, the practicalities of developing and estimating the human utility function, and how this agenda can assist in other methods of AI alignment. Listen here.

You can find all the FLI Podcasts here and all the AI Alignment Podcasts here. Or listen on SoundCloudiTunesGoogle Play, and Stitcher.

What We’ve Been Up to This Month

Max Tegmark & Meia Chita-Tegmark organized a workshop together with Emilia Javorsky on lethal autonomous weapons August 28-29, hosting an ideologically diverse group of world experts to see if they’d agree on more than a blank page. They produced a roadmap with fascinating ideas for first steps, including a 5-year use moratorium and strategies for verification, non-proliferation and de-escalation: https://www.cc.gatech.edu/ai/robot-lab/online-publications/AWS.pdf

Richard Mallah participated in the Partnership on AI’s All Partners Meeting in London.

FLI in the News

We’re new to instagram!
Please check out our profile and give us a follow:
@futureoflifeinstitute