MATS Summer 2026

The Summer 2026 program will run from June through August. It will be largest MATS program to date with 120 scholars and 100 mentors. Fellows will be connected with mentors or organizational research groups, such as Anthropic's Alignment Science team, UK AISIRedwood ResearchARC, and LawZero, to collaborate on a research project over the summer. Some fellows will be offered a 6+ month extension to continue this collaboration.

Program phases

Key dates for the application and admissions timeline

1. Applications

General Application (December 16th to January 18th) 

Applicants fill out a general application which should take 1-2 hours. Applications are due by January 18th.

Additional Evaluations (Late January through March)

Applicants that are advanced in the applications process go through additional evaluations including reference checks, coding tests, work tests, and interviews. Which evaluations you will undergo depend on the mentors and streams you apply to.

Admissions Decisions (Early April)
Selected applicants are notified of their acceptance and anticipated mentor later in the application cycle.

2. Main Program
3. Extension Phase
4. Post-program

Summer 2026 Streams

MATS supports researchers in a variety of research tracks, which includes technical governance, empirical, policy & strategy, theory, and compute governance. MATS fellows participate in a research stream consisting of their mentor(s) and other mentees. You can specify which tracks and streams to apply to in the general application. Each stream provides its own research agenda, methodology, and mentorship focus. 

SF Bay Area
Empirical
Security, Compute Infrastructure

Implementing SL4/5 and searching for differentially defense-favored security tools.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
SF Bay Area
Empirical
Dangerous Capability Evals, Adversarial Robustness, Security, Red-Teaming, Scalable Oversight

This stream will pursue research on securing and hardening AI systems through rigorous testing, provable defenses, and formal specification, including improving benchmarks for agentic security, scaling mathematically-grounded robustness techniques like randomized smoothing and Lipschitz-constrained training, and developing formal methods for specifying safe agent behaviors.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
Montreal
Empirical
Agent Foundations, Dangerous Capability Evals, Monitoring, Control, Red-Teaming, Scalable Oversight

We are excited to supervise projects that fall within the two following categories:

  1. Studying the causes, implications, and mitigations of [instances of] situational awareness;
  2. Contributing directly to LawZero's Scientist AI. 

For 1., we are particularly interested in:

  • Evaluation / monitorability awareness;
  • Self-awareness, in an introspective sense.

For 2., we are especially interested in:

  • Testing if  "truth-ification" (a process that, given a corpus of text, augments it so as to make sources of information explicit) allows language models to generalize better;
  • Developing amortized inference methods to estimate the uncertainty of a predictor (such as an autoregressive model). 
Mentorship structure
Desired scholar characteristics
Fellow project selection process
London
Empirical
Interpretability

Lee's stream will focus primarily on improving mechanistic interpretability methods for reverse-engineering neural networks.  

Mentorship structure
Desired scholar characteristics
Fellow project selection process
SF Bay Area
Technical Governance
Biorisk, Security, Safeguards

This stream will work on projects that empirically assess national security threats of AI misuse (CBRN terrorism and cyberattacks) and improve dangerous capability evaluations. Threat modeling applicants should have a skeptical mindset, enjoy case study work, and be strong written communicators. Eval applicants should be able and excited to help demonstrate concepts like sandbagging elicitation gaps in an AI misuse context.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
Tübingen
Empirical
Dangerous Capability Evals, Agent Foundations, Adversarial Robustness, Monitoring, Scalable Oversight, Scheming & Deception

Priority directions:

  • Risks from automating AI research
  • Automating safety and alignment research
  • AGI privacy
  • Measuring long-horizon agentic capabilities
  • New alignment methods
  • Science of post-training
Mentorship structure
Desired scholar characteristics
Fellow project selection process
London
Empirical
Control, Scheming & Deception, Dangerous Capability Evals, Monitoring

We will continue working on black-box monitors for scheming in complex agentic settings, building on the success of the previous stream.

See here for details.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
London
Empirical
Control, Scheming & Deception, Dangerous Capability Evals, Model Organisms, Monitoring

AI control focussed stream, probably running in-person in London.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
Washington, D.C.
Policy & Strategy
Policy & Governance, Strategy & Forecasting

Escalation risks from state perceptions of AI capability, AI-enabled targeting, AI-enabled decision manipulation, and the impact of AI integration into nuclear command and control. 

Mentorship structure
Desired scholar characteristics
Fellow project selection process
Washington, D.C.
Technical Governance
Compute Infrastructure, Policy & Governance, Security

This stream focuses on AI policy, especially technical governance topics. Tentative project options include: technical projects for verifying AI treaties, metascience for AI safety and governance, and proposals for tracking AI-caused job loss. Scholars can also propose their own projects.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
SF Bay Area
Empirical
Dangerous Capability Evals

This stream will focus on the science and development of model evaluations, especially monitorability and alignment evals.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
SF Bay Area
Technical Governance
Dangerous Capability Evals, Policy & Governance

Research papers (technical governance or ML) related to evaluating and mitigating dangerous AI capabilities, with a focus on what's actionable and relevant for AGI companies

Mentorship structure
Desired scholar characteristics
Fellow project selection process
SF Bay Area
Empirical
Security, Adversarial Robustness, Dangerous Capability Evals

This stream will focus on projects to better understand the capabilities of the model on dangerous capabilities specially more related to security. 

Also finding better ways to evaluate the safety and robustness of the models.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
SF Bay Area
Empirical
Dangerous Capability Evals, Red-Teaming, Model Organisms, Control, Monitoring

I'm interested in empirical projects that improve our ability to evaluate model capabilities or enable to understand or evaluate model monitorability. An ideal project culminates in a research output (conference/Arxiv paper or research blogpost with artifacts).

Mentorship structure
Desired scholar characteristics
Fellow project selection process
London
Empirical
Dangerous Capability Evals, Compute Infrastructure, Policy & Governance, Strategy & Forecasting

Making society safe from AI doesn't just mean making safe AI: we're figuring out how to uplift human collective intelligence, manage a highly multiagent world, improve foresight and institutional competence, ideally learning how to make best positive use of frontier AI systems as we go. FLF has a small, sharp team of researchers with a wide network, and we're looking to nurture new and missing approaches to minimising large-scale risks while steering to a flourishing future.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
Oxford
Theory
AI Welfare

Projects in this stream will be on AI welfare and moral status; more specifically, on what it takes to be a moral patient and how we can determine whether AI systems meet the conditions. I'm looking for applicants who have ideas about these topics and are motivated to explore them in more detail.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
SF Bay Area
Empirical
Interpretability

In this stream we will explore extensions and implications of our discovery that neural networks pretrained on next-token prediction represent belief-state geometry in their activations. We will build on this fundamental theory of neural network representations in order to discover what AI systems are thinking, and understand their emergent behaviors.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
New York City
Technical Governance
Dangerous Capability Evals, Control, Strategy & Forecasting, Policy & Governance, Scalable Oversight, Agent Foundations

Peter Henderson’s stream focuses on developing safe, aligned AI agents, with projects on scalable oversight rules informed by law and game theory, safe long-horizon exploration, and measuring “jagged” capability/safety frontiers. Scholars will join an independently driven, engineering-heavy research environment, collaborating with other MATS scholars and PhD students, with weekly 1:1s and active async mentorship.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
SF Bay Area
Empirical
Control, Model Organisms, Scheming & Deception, Strategy & Forecasting

The Redwood Research stream is looking for fast empirical iterators and strategists to work on control research.

Mentorship structure
Desired scholar characteristics
Fellow project selection process
SF Bay Area
Theory
Agent Foundations

My MATS fellows will do philosophical thinking about multi-agent intelligence and how agents change their values. This will likely involve trying to explore and synthesize ideas from game theory, signaling theory, reinforcement learning, and other related domains.

Mentorship structure
Desired scholar characteristics
Fellow project selection process

Related research

No items found.

Community at MATS

MATS Research phase provides scholars with a community of peers.

Scholars work out of a shared office and are supported by the Community Team.

MATS alumni report that the connections with peers that they made during MATS have had the largest impact on them years later. Our full-time Community Team works to facilitate these connections and also provide general well-being support. Weekly lightning talks, scholar-led discussion groups, game nights, and outings to SF are some examples of MATS events.