Fazl Barez
How do we make AI systems that we can actually inspect, correct, and trust? Fazl Barez is a Principal Investigator at the University of Oxford leading research on technical AI safety, interpretability, and governance. He teaches Oxford’s AI Safety and Alignment course, has consulted for Anthropic, and held fellowships at RAND Corporation. His work bridges mechanistic interpretability, model auditing, and policy — translating what’s found inside models into tools regulators and auditors can use.