
Serena Booth
Why do you care about AI Existential Safety?
AI systems have potential to be beneficial to humanity, but the risks these systems pose are equally tremendous. These potential outcomes are key to my research.
Please give at least one example of your research interests related to AI existential safety:
I work on the problem of specification design, which is central to AI safety. Namely, how can we empower people to write correct specifications? What modalities—like preferences, corrections, explanations—can help with this endeavor? And, after writing a specification, how can a person know that the AI system has learned the intended interpretation of the specification?