Dylan Hadfield-Menell
Why do you care about AI Existential Safety?
With AI systems, you often get what you can measure. This creates a structural bias towards simpler measures of value and runs the risk of diverting more and more resources towards these simple goals. My interest in existential safety comes from a desire to make sure that technology supports and nurtures a rich and diverse set of values.
Please give one or more examples of research interests relevant to AI existential safety:
I work on the theoretical and practical study of machine alignment. This includes: methods for value learning from observations; algorithms to optimize uncertain objectives; formal analysis of design/oversight strategies for AI systems; as well as the study of incomplete goal specifications and corresponding consequences of overoptimization.