
UK AISI
—
Research Scientist
Sid Black is a Research Scientist at UK's AI Safety Institute building autonomy evaluations. He previously co-founded EleutherAI and Conjecture, where his work focused on pretraining and evaluating large language models, interpretability, and programming agents.
His current research interests include:
- understanding and quantifying autonomy risks from AI agents.
- Identifying relevant autonomy threat models.
- Building new evaluations which increase our coverage of these threat models.
- Researching ways to increase resilience to AI risks, in the case of unknown or unpredictable threat models.
- Developing standards for, and the science of, AI agent evaluations.
- Predictive evaluations and White-box evaluations for AI agents.