Skip to content
All Newsletters

FLI March, 2016 Newsletter

Published
14 March, 2016
Author
admin

Contents

FLI March, 2016 Newsletter
The first two months of 2016 were full of news, activities, conferences, talks, and more…

The first two months of 2016 were full of news, activities, conferences, talks, and more…

The Growing Risk of Nuclear Weapons

Nuclear weapons have been around for over half a century, but many experts believe they’re becoming a bigger risk. To emphasize just how big a threat nuclear weapons are, we’ve created a timeline of close calls, any of which could have accidentally triggered a nuclear war. We’ve also worked closely with David Wright at the Union of Concerned Scientists to create two podcasts — one about the general nuclear threat and one about North Korea’s nuclear test — and he also wrote two blogs about the North Korean satellite launch.
Save the date: On April 2 we’re helping to organize a nuclear conference featuring former Secretary of Defense William Perry and Cambridge Mayor Denise Simmons, along with many of the distinguished participants. More information to come. Perry recently warned about the risks of  nuclear weapons at Google, as well as in his book, My Journey at the Nuclear Brink.

AAAI: Featuring FLI Grantees at the AI Safety Workshop

The AAAI conference was held in Phoenix, AZ this February, and FLI was honored to help organize and participate in the AI, Ethics, and Safety Workshop. The afternoon session featured talks by many of the recipients of the FLI AI safety grants that were awarded in July of last year, and it also included a panel discussion about what types of research projects should be considered next. The conference officially kicked off with a speech by President Tom Dietterich about the importance of robust AI research. Stuart Russell and Peter Norvig received the Outstanding Educator Award. And Nick Bostrom gave a plenary talk about the importance of AI safety and the future of AI. He and FLI advisor Erik Brynjolfsson also participated in a panel discussion about the impacts of AI on jobs.

The Future of AI

In January, Yann LeCun hosted a symposium about what the future of artificial intelligence will look like. FLI’s Max Tegmark spoke at the symposium, along with leaders in the tech industry and AI research community, including Eric Schmidt of Alphabet, Demis Hassabis of Deep Mind, Eric Horvitz of Microsoft, Jen-Hsun Huang of NVIDIA, and many others. While the discussion was mostly optimistic, much of it also focused on keeping AI safe and beneficial, which is a testament to just how far we’ve come since FLI hosted its Puerto Rico conference last January.

AlphaGo

DeepMind made big news  with an announcement that their program, AlphaGo, had defeated the Go European Champion. Some influential AI researchers gave us their take on the breakthrough.
Don’t forget: AlphaGo will play the reigning world champion, Lee Sedol, on March 9-15. You can watch it live on Youtube.

Metaculus

Predicting the future (of life): FLI co-founder, Anthony Aguirre, recently launched an exciting, online predictive tool, Metaculus. Metaculus is a great place to ask questions about the future and make your own predictive analysis. Not sure where to start? Have a go with the AlphaGo predictions  for the upcoming games!

Intro to AI Safety

Good AI safety resources can get lost among the ever growing sea of new AI research and news articles that may or may not be well informed. FLI co-founder, Victoria Krakovna, put together an excellent starter list of resources to help people new to AI safety understand what the issues and research are about.

Who’s to Blame?

Matt Scherer, lawyer and AI enthusiast, started blogging for us with a series called “Who’s to Blame,” which considers just how unprepared the legal realm is for autonomous weapons. His posts include:

Is the Legal World Ready for AI?
The Legal Vacuum Surrounding Autonomous Weapons
What Is an “Autonomous” Weapon?
Could Autonomous Weapons Systems Navigate the Law of Armed Conflict?
Who’s to Blame if an Autonomous Weapon Breaks the Law?

News of the Week

We’ve recently begun a new weekly feature to highlight the most interesting and/or relevant news pertaining to x-risk and x-hope. The news we thought was most important:

Human Embryo Gene Editing
Nuclear Winter and a Government Risk Report
AAAI, Beneficial AI Research, a $5M Contest, and Nuclear Risks
Ocean Warming and Nuclear Protests

CRISPR

CRISPR is THE big news in the biotechnology field these days, from gene drives to wipe out mosquitos to patent battles to editing human embryos, CRISPR is in headline after headline. We highlighted the most exciting CRISPR developments of the last few months to bring you up to speed.

Our newsletter

Regular updates about the Future of Life Institute, in your inbox

Subscribe to our newsletter and join over 20,000+ people who believe in our mission to preserve the future of life.

Recent newsletters

Future of Life Institute Newsletter: Tool AI > Uncontrollable AGI

Max Tegmark on AGI vs. Tool AI; magazine covers from a future with superintelligence; join our new digital experience as a beta tester; and more.
2 December, 2024

Future of Life Institute Newsletter: Illustrating Superintelligence

Need a break from US election news? Explore the results of our $70K creative contest; new national security AI guidance from the White House; polling teens on AI; and much more.
1 November, 2024

Future of Life Institute Newsletter: On SB 1047, Gov. Newsom Caves to Big Tech

A disappointing outcome for the AI safety bill, updates from UNGA, our $1.5 million grant for global risk convergence research, and more.
1 October, 2024
All Newsletters

Sign up for the Future of Life Institute newsletter

Join 40,000+ others receiving periodic updates on our work and cause areas.
cloudmagnifiercrossarrow-up linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram