
Samuel Albanie
Position
Assistant Professor
Organisation
University of Cambridge
Member of
Biography
Why do you care about AI Existential Safety?
As technology grows more powerful, the consequences of failure for that technology also grow. Over the last few decades, significant advances in research across both hardware and software have yielded meaningful gains in AI capabilities. In the absence of hard limits imposed by physics that preclude further gains, I believe it is prudent to consider the question of what happens if progress continues and to allocate research effort towards mitigating the risks of this eventuality.
Please give at least one example of your research interests related to AI existential safety:
- Foundation models have yielded striking gains across a broad suite of tasks spanning text, vision and code. However, the self-supervised pretraining objectives of these models often do not precisely match the desired end objective of the user. Given this disparity, I'm interested in the question of how these models can be induced to be maximally helpful for human end users. Currently, I'm considering this problem through the lens of natural language prompting - communicating tasks through utterances.
- I'm also interested in better understanding the potential for machine learning models to learn manipulation strategies (see e.g. https://arxiv.org/abs/1701.04895)
Our content
Content from this author
Sort order
Category
Content type
Number of results
May 22, 2023
Halogen-catalyzed reactions on smoke destroy the ozone layer after regional nuclear war
grant
May 22, 2023
grant
May 22, 2023
grant
May 22, 2023
Improving the representation of crop production losses due to nuclear conflict (CODEC)
grant
May 22, 2023
grant
May 22, 2023
The cascading impacts of postnuclear ultraviolet radiation on photosynthesizers in the Earth system
grant
May 22, 2023
grant
May 22, 2023
WUDAPT-based framework for numerical simulations of nuclear urban fires and pyroconvective plumes
grant
May 22, 2023
grant
May 22, 2023
grant
May 22, 2023
Advanced ensemble projections for indirect impacts of nuclear war in global food systems
grant
May 22, 2023
grant
May 4, 2023
Nathan Labenz on the Cognitive Revolution, Red Teaming GPT-4, and Potential Dangers of AI
podcast
May 4, 2023
podcast