
Truthful AI
—
Research Scientist
Owain has a broad interest in AI alignment and reducing AGI risk. He is investigating dangerous capabilities and the emergence of misalignment in LLMs, along with self-awareness and latent reasoning. Owain previously worked on AI deception (How to Catch an AI Liar), truthfulness (TruthfulQA), and the Reversal Curse. Owain runs an independent AI Safety non-profit, based at Constellation in Berkeley. He previously worked at the University of Oxford and at Ought. He has mentored 30+ junior AI Safety researchers through MATS and other programs.