Serge Egelman | 2024 Norm Hardy Prize Seminar

Abstract: Internet users often neglect important security actions (e.g., installing security updates or changing pass-words) because they interrupt usersā€™ main task at inopportune times. Commitment devices, such as reminders and promises, have been found to be effective at reducing procrastination in other domains. In a series of on-line experiments, we explored the effects of reminders… Continue reading Serge Egelman | 2024 Norm Hardy Prize Seminar

2025 Fellow Maxx Yung

2025 Fellow Max Langenkamp

2025 Fellow Gavin Leech

2025 Fellow Felix Sosa

2025 Fellow Daniele Palombi

2025 Fellow Abhinav Singh

R. Pihlakas | Biologically and Economically Aligned Multi-objective Multi-agent AI Safety Benchmarks

Presenter Roland Pihlakas Roland Pihlakas, MSc in psychology, is an experienced AI software engineer who has been working on multi-objective value problems for almost 20 years, and has followed discussions on AI Safety since 2006. His thesis topic was about modelling of innate learning and planning mechanisms, which eventually form a foundation for culturally acquired… Continue reading R. Pihlakas | Biologically and Economically Aligned Multi-objective Multi-agent AI Safety Benchmarks

Dmitrii Usynin, TU Munich | Meaningful evaluation of Privacy-Preserving Machine Learning

Marc Carauleanu | Self-Other Overlap: A Neglected Path to Existential Safety

Presenter Marc Carauleanu Marc Carauleanu is an AI Safety Researcher at AE Studio. They are leading a team investigating a neglected AI Safety proposal that is focusing on self-other overlap: the AI model having similar internal representations when it reasons about others to when it reasons about itself. This agenda was heavily inspired by scientific… Continue reading Marc Carauleanu | Self-Other Overlap: A Neglected Path to Existential Safety

0
    0
    Your Cart
    Your cart is emptyReturn to Shop