Projects

Research Direction: Automated Interpretability

  • AI Agents for Interpretability

  • Training Models to explain their own computations

  • Empowering Users through Explainability

What I'm looking for in a Mentee

  • Strong technical background in AI and ML

  • Good programming skills

  • previous experience in ai safety/interpretability would be a plus

  • Independent and motivated

  • Quick learner

Bio

Fazl Barez is a Senior Research Fellow at the University of Oxford dedicated to building AI that strengthens human agency. His work spans interpretability, safety, unlearning, and governance, developing both technical systems and institutional safeguards that allow people to understand, shape, and direct powerful AI models. His research is motivated by a simple principle: AI should amplify human judgment and autonomy — not replace it, obscure it, or centralise it in the hands of a few.

Next
Next

Jérémy Scheurer