Skip to content

He He

Position
Assistant Professor
Organisation
New York University
Biography

Why do you care about AI Existential Safety?

As AI systems become more competent and deployed into critical social sectors, it is concerning that their profound impact on society is often studied in a post-hoc way (e.g., influence on election, social polarization). While it is hard to predict the future trajectory (several more paradigm shifts might need to happen before we reach general or strong AI), I think “improving human wellbeing” should be the central objective from the very beginning when we design AI systems.

Please give one or more examples of research interests relevant to AI existential safety:

My current research interest lies in trustworthy AI, with a focus on natural language technologies. To make reliable and safe decisions, the learning system must avoid catastrophic failure when facing unfamiliar (out-of-distribution) scenarios. We aim to understand what types of distribution shifts incur risk and how to mitigate them. We are also excited by the prospect of AI systems collaborating with and learning from their human partners through natural language interaction. To this end, our work has focused on factual text generation models (that do not lie) and collaborative dialogue agents.

Sign up for the Future of Life Institute newsletter

Join 40,000+ others receiving periodic updates on our work and cause areas.
cloudmagnifiercrossarrow-up linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram