Skip to content

Dylan Hadfield-Menell

Assistant Professor
Massachusetts Institute of Technology

Why do you care about AI Existential Safety?

With AI systems, you often get what you can measure. This creates a structural bias towards simpler measures of value and runs the risk of diverting more and more resources towards these simple goals. My interest in existential safety comes from a desire to make sure that technology supports and nurtures a rich and diverse set of values.

Please give one or more examples of research interests relevant to AI existential safety:

I work on the theoretical and practical study of machine alignment. This includes: methods for value learning from observations; algorithms to optimize uncertain objectives; formal analysis of design/oversight strategies for AI systems; as well as the study of incomplete goal specifications and corresponding consequences of overoptimization.

Sign up for the Future of Life Institute newsletter

Join 40,000+ others receiving periodic updates on our work and cause areas.
cloudmagnifiercrossarrow-up linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram