AI safety discourse often splits into immediate harm vs catastrophic risk framings. In this keynote, I argue that the two research streams will benefit from increased cross-talk and a greater number of synergistic projects. A zero-sum framing on attention and resources between the two communities is incorrect and does not serve either side's goals. Recent theoretical work, including on accumulative existential risk, unifies risk pathways between the two fields. Building on this, I suggest concrete synergies that are already in place - as well as opportunities for future collaboration. I will discuss how shared research and monitoring infrastructure, such as UK AISI Inspect, can benefit both areas; how methodological approaches from human behavioral science, currently used in immediate harms research, can be ported into AI behavioral science applied to existential risk research; and how technical solutions from catastrophic risk research can be applied to mitigate immediate societal harms. We have a shared goal of building a better, safer future for everyone. Let's work together!
talk-data.com
K
Speaker
Katarina Slama, PhD
1
talks
Research Scientist
UK AI Security Institute
Katarina has a PhD in Neuroscience from UC Berkeley, focusing on mechanisms of human attention allocation. Previously, She worked as a Member of Technical Staff at OpenAI, where she contributed to the development of InstructGPT in the pre-ChatGPT era, and later worked on evaluations of societally relevant model behaviors. Katarina is broadly interested in the societal impacts of Transformative AI, especially AI impacts on mental health. Based in London, she is quickly leveling up on my appreciation for Yorkshire pudding and shepherd's pie.
Bio from: Virtual Keynote Talk "AI Safety: Near and Far"
Filter by Event / Source
Talks & appearances
1 activities · Newest first