
Anthropic
Dan Valentine is a Member of Technical Staff at Anthropic, an AI safety and research company. His work is primarily focused on AI safety and alignment research, including scalable oversight methods and understanding how AI models interact with data and prompts.
The Summer 2023 cohort supported 60 scholars with 15 mentors, working across 12 different research areas. The program consisted of a remote 4-week training phase, an 8-week research phase in Berkeley, and a 4-month extension phase. MATS leadership co-founded the London Initiative for Safe AI (LISA) in September 2023 to provide a dedicated research space for AI safety researchers and organizations in London, and for MATS scholars to continue their research projects. Research projects were distributed across multiple areas, with approximately one-third focused on evaluations and capability demonstrations and one-fifth on mechanistic interpretability, alongside work on agent foundations, activation engineering, and cooperative AI.
Ethan spent a lot of time discussing our research with us and gave great advice on direction. He unblocked us in various ways, such as getting access to more models or to lots of compute budget. He connected us with lots of great people, some of whom became collaborators. And he was a very inspiring mentor to work with.
Dan Valentine