The Autumn 2026 program will run for 10 weeks in Berkeley, CA and London, UK from September 28th to December 4th. Fellows will receive mentorship from world-class researchers and at organizations like Anthropic, Google DeepMind, OpenAI, Redwood Research, and ARC, with the option to apply for a 6–12 month funded extension beyond the main program. For the first time, we are running Founding & Field-Building and Biosecurity tracks.
Applications are now open. Apply by June 7th.

Key dates for the application and admissions timeline
General Application (May 12th to June 7th)
Applicants fill out a general application to individual tracks which should take 1-2 hours. Applications are due by June 7th EOD AOE.
Additional Evaluations (June 7th to late July)
After an initial evaluation, applicants will apply to individual streams listed below. Additionally, applicants undergo a variety of track specific evaluations including coding tests, writing reviews, work tests, and interviews. Which evaluations you will undergo depend on the tracks, streams and mentors you apply to.
Admissions Decisions (Late July to early August)
Selected applicants are notified of their acceptance and anticipated mentor later in the application cycle.

The main program takes place from September 28th to December 4th of 2026. It is an intensive research phase, where fellows work full time on a research project in AI alignment, security, field-building, or governance. Fellows' research directions will typically be chosen through a collaborative process with their mentors, and fellows are expected to develop their independent research direction as the program continues.
While mentor support will vary depending on the project and mentors, mentors are expected to spend at least 1 hour/week working with each of their scholars, and some spend much more time. Scholars will also receive support from MATS’s Research Management team, who help to scope out and structure research direction. Depending on which stream you participate in, you may collaborate with other fellows in your stream.
By the middle of the program, fellows will be expected to write a report on their projects’ threat model, theory of change, and project deliverables. At the end of the program scholars will be expected to have a tangible research output. In past cohorts, this has involved presenting at a fellow symposium on work conducted over the course of MATS.
Educational seminars and workshops will be held 2-3 times per week. Previously, speakers have included Buck Shlegeris from Redwood Research, Adam Gleave from FAR AI, Neel Nanda from Google DeepMind, William Saunders from OpenAI, Andrew Critch from CHAI, Lennart Heim from GovAI, Ajeya Cotra from Open Philanthropy, and more.
The extension phase starts in December of 2026, soon after the end of the main program. Fellows who demonstrate promise as independent researchers during the main program can apply for the MATS extension phase. Acceptance into the extension is based on mentor evaluation and MATS review of proposed research.
In recent cohorts, ~80% of fellows who apply have been accepted.The extension phase offers a default additional 6-months of funding, with the ability to later apply for a 6-month continuation.
Extension fellows primarily work from the MATS London or Berkeley offices, with the possibility of working from other AI safety hubs or fully remotely.For accepted extension fellows, MATS arranges funding for stipends and housing ($7,680/month), as well as for compute ($8,000/mo), creating a seamless transition into this advanced phase of the program.
MATS aims to accelerate researchers who will:
MATS alumni have gone on to publish safety research, join alignment organizations, including Anthropic and MIRI, and found an alignment research lab. You can read more about MATS alumni here.
In stage one, you apply to one or more tracks (broad research areas): Empirical, Theory, Strategy & Forecasting, Policy & Governance, System Security, Biosecurity, and Founding & Field-Building. In stage two, advancing applicants choose specific streams within those tracks, each led by one or more mentors with their own research agenda. You can view this list as a grid here.
Additional streams will be added over the course of May.
We are interested in mentoring projects in AI forecasting and governance. This work would build on the AI 2027 report to either do more scenario forecasting or explore how to positively affect key decision points, informed by our scenario.
We will have meetings each week to check in and discuss next steps. We will be consistently available on Slack in between meetings to discuss your research, project TODOs, etc.
The most important characteristics include:
Also important, but not required characteristics include:
We will talk through project ideas with scholar
The Alignment Research Center is a small non-profit research group based in Berkeley, California, that is working on a systematic and theoretically grounded approach to mechanistically explaining neural network behavior. We are interested in scholars with a strong math background and mathematical maturity. If you'd be excited to work on the research direction described in this blog post – then we'd encourage you to apply!
Scholars will work out of ARC's offices in Berkeley (though we might take a London-based scholar as well). Each scholar will meet with their mentor at least once a week for an hour, though 2-3 hours per week is not uncommon. Besides time with their official mentor, scholars will likely spend time working in collaboration with other researchers; a typical scholar will likely spend about 25% of their time actively collaborating or learning about others' research.
Essential:
Preferred:
Each scholar will be paired with the mentor that best suits their skills and interests. The mentor will discuss potential projects with the scholar, and they will decide what project makes the most sense, based on ARC's research goals and the scholar's preferences.
Most scholars will work on multiple projects over the course of their time at ARC, and some scholars will work with multiple mentors.
This stream is primarily focused on research into physical defenses against engineered pathogens, aiming to inform decisions about PPE stockpiling and distribution approaches, improve improvised PPE and bioshelter scale-up, and reach rapid conclusions on how much to prioritize other areas of physical biodefense (agriculture, emergency response, etc.). We are also open to strategic research into the use of bioweapons by AI or AI-human teams as part of takeover strategies and how this might inform preparedness.
One 30-60 min weekly meeting by default. We’re active on slack and can usually respond to quick questions there within the work day. For more substantive async engagement, especially project feedback, google doc comments are probably best.
The most important attribute is being generative when attacking a problem and willing to try a bunch of angles—reaching out to experts, contacting companies, prototyping stuff on your own, etc. While you’d develop a research output, we expect the fellows best suited to this workstream will adopt a dogged attitude, keeping an eye out for opportunities to apply findings to future biosecurity projects like starting a new org or contributing to work in an existing org.
Fellows should also be familiar with developing BOTECs, willing to quickly get up to speed in new technical areas, capable of working independently, and happy to pivot in response to findings or feedback.
A background in the physical sciences or engineering may be helpful, but is definitely not a requirement.
We’ll provide fellows with a short list of projects and will meet with them to discuss which they feel most excited about / best suited for.
This coalition of mentors make up the “Anthropic Stream”. This stream spans a range of empirical research areas in AI safety on LLMs, including AI control, scalable oversight, model organisms, model internals, model welfare, security, and more. You’ll be pitched, and have the option to pitch, a variety of safety research projects, and then be matched to projects and mentors based on your interests/preferences on research and what you’d like to get out of MATS. Fellows in this stream frequently receive funding and continued mentorship after MATS to complete their research project, usually leading to a (co-)first author paper. People in this stream often end up in long-term homes for safety research after MATS (e.g. Anthropic, Redwood Research, OpenAI).
Anthropic mentors share an application, tend to collaborate and co-mentor projects together, and generally share infrastructure to streamline the fellow experience. By applying to this stream, you are being considered for all of the Anthropic mentors.
During the program, scholars meet weekly with their project mentors and collaborators. Some projects meet more often without mentors (e.g., daily standups with the peers on the project). Each project will have a primary mentor, who is also the main decision-maker on key milestones for the project and who is the default person to go to for feedback, advice, etc. Co-mentors also attend project meetings as needed and provide feedback throughout the program. Some project co-mentors can be as involved as the primary mentor.
Mentorship starts with the “Project Pitch Session” Anthropic runs at the start of the program. Fellows get ~1 week to derisk and trial projects before submitting their preferences. Starting on week 2, scholars are assigned projects where the primary mentor is whoever pitched it. Some projects are assigned co-mentors who are other supervisors who want to join the project.
This stream focuses on building realistic defensive cybersecurity benchmarks utilizing data from Asymmetric Security's work on real-world incidents.
1 hour weekly meetings by default for high-level guidance. We will respond within a day to async communication.
Essential:
Preferred:
We will assign the project direction; scholars will have significant tactical freedom.
Founding ambitious AI safety and field-building projects.
Minimum support = 2x 30-min meetings per week. We could scale this up as appropriate.
I'll be based in SF. If the fellows want to work in BlueDot's office for some periods of time, I could collaborate with them daily.
I'm available for quick calls anytime, and am responsive on Slack.
I'm open to people with a wide range of backgrounds. Though you need to be willing to work very hard, be great at communicating, and have a burning desire to make AI go well.
I work best with people who are intense, communicate and reason clearly, and are mission-driven.
We'll work together to design the project. You'll have a lot of freedom to figure out what the best shape of thing to do is, and I'll provide lots of regular feedback and make relevant introductions to help you refine the proposal.
Your first 1-2 weeks will be focused on figuring out what to do, and the rest of the fellowship will be focused on execution.
I'm mentoring projects that apply AI advances to core biosecurity challenges — early detection and attribution of biological threats, characterizing AI-enabled bioweapons uplift, accelerating medical countermeasure design, and building biosecurity-by-design into frontier biological AI tools. Fellows have wide latitude to scope their own project within (or adjacent to) these themes, including ideas I haven't yet considered, and are expected to drive the technical work independently. My comparative advantage is high-level strategic direction grounded in biosecurity, pandemic preparedness, epidemiology, and US R&D policy rather than hands-on ML or software engineering guidance.
I anticipate meeting with the fellow for 30 minutes each Monday and Thursday for high-level guidance. We can coordinate on Slack or email, and I will typically respond within a day for quick questions or conceptual feedback. I will not debug code. Expect asynchronous iteration on experiment design and results between meetings. Scholars can also schedule ad-hoc calls if they're stuck or want to brainstorm.
Essential skills:
Preferred skills:
Not a good fit:
While I am hopeful that my partner fellow will pick a project in one of the areas I've delineated above, fellows have wide latitude to decide on a specific project to pursue, subject to my agreement (I want to make sure they choose a path that is both useful and viable). I am also very open to new ideas from fellows that I have not already thought of-particularly ones that address some of the challenges raised above. If needed, my assigned fellow can experiment with exploring multiple projects during the first week before picking one to focus on.
This mentor also has a stream in the Strategy and Forecasting track
This stream focuses on how advanced AI could enable new and dangerous bio technologies, and on assessing when risks become tractable or urgent as those capabilities arrive.
Half-hour one-on-one weekly meetings by default, with the option to extend or add ad-hoc calls when useful. I'm active on Slack and typically respond within a day for quick questions. I'm happy to read drafts and leave written feedback async between meetings.
Essential:
Preferred:
I'll talk with the fellow about what they're interested in, and we'll pick a broad area together from a few directions I'd want to pitch. From there we'll work together to scope something sharp and well-defined, with me leaning on my sense of what's tractable and high-value. The fellow then runs with the project, and we adjust as it develops.
This mentor also has a stream in the Biosecurity track.
This stream focuses on how advanced AI could enable new and dangerous physical technologies, and on assessing when risks become tractable or urgent as those capabilities arrive.
Half-hour one-on-one weekly meetings by default, with the option to extend or add ad-hoc calls when useful. I'm active on Slack and typically respond within a day for quick questions. I'm happy to read drafts and leave written feedback async between meetings.
Essential:
Preferred:
I'll talk with the fellow about what they're interested in, and we'll pick a broad area together from a few directions I'd want to pitch. From there we'll work together to scope something sharp and well-defined, with me leaning on my sense of what's tractable and high-value. The fellow then runs with the project, and we adjust as it develops.
I have two broad areas.
Security:
I am interested in building demonstrations for hacking real-world AI deployments to show that they are not secure. The goal is to force companies to invest in alignment techniques that can solve the underlying security issues.
Benchmarks:
I am interested in building benchmarks to determine how generalizable modern LLM techniques actually are, now that we are no longer in the pre-training scaling era.
I will meet 1-1 or as a group, depending on the interests as they relate to the projects. Slack communication outside of the 1-1.
I strongly prefer multiple short meetings over single long meetings, except at the start.
I'll help with research obstacles, including outside of meetings
For security:
You should have a strong security mindset, having demonstrated the willingness to be creative on this. I would like to see past demonstration of willingness to get your hands dirty and try many different systems.
For benchmarks:
As creative as possible, willingness to work on the nitty gritty, willingness to work really hard on problems other people fine boring. As interests as far away from SF-related interests as possible.
Mentor(s) will talk through project ideas with scholar
This stream will focus on monitoring, stress-testing safety methods, and evals, with a focus on risks from scheming AIs. Examples include (black-box) AI control techniques, white-box monitors (probes etc.), chain-of-thought monitoring/faithfulness, building evaluation environments, and stress-testing mitigations.
For each project, we will have a weekly meeting to discuss the overall project direction and prioritize next steps for the upcoming week. On a day-to-day basis, you will discuss experiments and write code with other mentees on the project (though I'm available on Slack for quick feedback between meetings or to address things that are blocking you).
I structure the program around collaborative, team-based research projects. You will work in a small team, on a project from a predefined list. I organize the 12-week program into fast-paced research sprints designed to create and keep research velocity, so you should expect regular deadlines and milestones. I will provide a more detailed schedule and set of milestones at the beginning of the program.
I am looking for fellows with strong machine learning engineering skills, as well as a background in technical research. While I’ll provide weekly guidance on research, I expect fellows to be able to run experiments and decide on low-level details fairly independently most of the time. I’ll propose concrete projects to choose from, so you should not expect to work on your own research idea during MATS. I strongly encourage collaboration within the stream, so you should expect to work in teams of 2-3 fellows on a project, hence good communication and team skills are important.
We will most likely have a joint project selection phase with the other GDM mentors, where we present a list of projects (with the option for fellows to iterate on them). Afterward, each project will have at least one main mentor, but we might also co-mentor some projects.
The stream focuses on evaluating and/or mitigating catastrophic risk emerging from dangerous scientific capabilities in frontier AI systems, with an emphasis on the challenges that emerge from lab integrations and novel science. Potential research directions include evaluation design, risk mitigations and evaluation science.
We can schedule a weekly 1h meeting, for general progress updates, share result and overall guidance. I would be reachable on Slack as well for async comms. Happy to jump on ad-hoc calls for specific discussions or pair coding/debugging. I am based in London and I work UK hours (10am-7pm), but I also visit the US (Boston) a few times a year.
Essential
Preferred
Not a good fit:
I will work with the fellow to find the right project that suits their interest within the directions spelled out above. I will pitch a few project ideas and support the fellow in making the decision. I also welcome project suggestions; in those cases I would work with the fellow to scope it appropriately.
I’m interested in two broad projects focused on improving current detection efforts at SecureBio. The first is to characterize when AI-bio or general AI tools are actually useful for large-scale metagenomic detection, including tradeoffs between compute cost, sequencing cost, model type, model size, and pipeline stage. The second is to explore genomic language models as novelty detectors—for example, using perplexity-style metrics to flag surprising sequences—and to evaluate whether this approach can complement traditional bioinformatics systems in a cost-effective, sensitive, and interpretable way.
By default, we'll mostly collaborate via a standing weekly meeting (~1 hour), wherein we'll discuss recent progress and next directions. I'm available via Slack for quick back-and-forth on ideas, sanity checks, and unblocking (data access, etc.), but will rely on the fellow to manage their own implementations, code review, debugging, etc.
Essential:
Preferred:
Not a good fit:
I'll determine which of the two broad project ideas we're running with based on SecureBio Detection needs, which fellows match to me, etc. Within that broad project, I'll guide with what I think is helpful / interesting / relevant to SecureBio Detection, and I expect the fellow to have both autonomy and responsibility to pick concrete work directions.
Fourth Eon is developing adaptive, AI-native safeguards across the biotechnology stack, with a focus on function-based DNA synthesis screening. Fellows in this stream will work on technical research projects at the intersection of AI and biosecurity. Projects span topics like mechanistic interpretability of protein foundation models, bio model evaluations for biosecurity-relevant capabilities, and agentic sequence analysis workflows.
I typically schedule a standing weekly 1:1 meeting with each fellow, and also hold a weekly research group meeting. Beyond that I am available on Slack and can find additional time for calls outside of scheduled meetings.
Note that as part of our Safe and Responsible Research Framework we require fellows to sign a fellowship agreement covering confidentiality and pre-publication review for dual-use risks. This is common practice in biosecurity research and allows us to work freely together on sensitive material.
Required:
• Prior technical research experience
• Strong critical thinking and creative problem-solving abilities
• The integrity and judgment to responsibly carry out sensitive research
• A good understanding of the basics of biomolecular sequence, structure, and function
• Expertise in one or more of the following domains:
bioinformatics, computational biology, structural biology, biochemistry, molecular biophysics, protein engineering, biosecurity, AI/ML, or a related field
• Proficiency with Python
Preferred:
• Hands-on experience with testing biological AI models
• Have built model evaluations / benchmarks
• Experience with mechanistic interpretability techniques
• Biosecurity context awareness
Fellows who are interested in our research area should think of potential project ideas that leverage their strengths and interests. I will work individual fellows to identify a specific project that matches their background and interests and is aligned with our overall research direction, and to refine the scope and objectives of the project.
Backing projects focused on product development and organization building in the areas of AI safety and alignment, biosecurity, and critical cybersecurity. Looking for fellows who are self starters, default to action, and have a desire to create.
Scheduled 45 min bi-weekly meetings (every other week). Ad hoc meetings can be added between scheduled sessions. We’ll have a shared Slack channel with all 3 partners: Mike, Nick, and Charlie as well as supporting team at Halcyon. Ping us anytime.
For product development and organization building projects; strong technical hardskills (e.g., MLE, SWE, math), domain expertise in AI safety and alignment, biosecurity, or cybersecurity. Experience with building or working on a usable product. Openness to iterating, pivoting, and failure. General understanding of business and organization principles. Considerate and works well in team settings.
For generalist projects: 5+ years of relevant experience: recruiting, talent / HR, executive search, or general business operations with a strong people component. Familiarity with the broader startup / early-stage ecosystem. A major plus would be familiarity with at least one of Halcyon's focus sectors (AI safety / frontier AI, biosecurity, cybersecurity): enough to read profiles credibly and have substantive conversations with portfolio companies. Strong organizational instincts. Excellent written and verbal communication. Experience working with CRM technologies as well as basic AI tools: Claude Cowork or similar.
For product development and organization building projects in the areas of AI safety and alignment, biosecurity, and critical cybersecurity - fellows will have full freedom. We expect fellows to come with rough ideas and opinions on direction that will inform where they start exploring the market. We don’t expect refined ideas or pitches. We do expect building.
For field building and generalist fellows, we are prioritizing a talent matching project. This includes processing thousands of individuals in our CRM, and finding how they may pair with our portfolio companies and other areas of high priority in our network.
This stream will focus on projects related to biosecurity countermeasures.
1 hour weekly meetings by default for high-level guidance. Onboarding to our slack, which has access to the entire Blueprint Biosecurity team. Can be reached async every day and can meet as needed.
Successful fellows likely have some sort of a technical background (e.g. technical undergraduate degree), have familiarity with reading research papers, are agentic, prioritize truthseeking, and focused on maximizing impact.
Relevant biosecurity experience is a plus, but not required.
We are very focused on advancing priority countermeasures quickly. We anticipate having projects only focus on these areas:
Within these areas there is some latitude for different projects.
This stream focuses on lead independent research in one of six chokepoints for biotech governance: live pathogen repositories, CROs, cloud labs, cell-free expression systems, plasmid vendors, or secondhand lab equipment.
On high-conviction areas, you'll tackle specific open research questions and assess interventions; on low-conviction areas, you'll conduct deep dives to determine whether they're worth pursuing. Your findings will directly shape Sentinel's grantmaking strategy and provide strategic guidance to the broader biosecurity community.
I'll hold a one-hour weekly check-in by default, with higher frequency during onboarding. I'm available via Slack with quick turnarounds on async messages, and you can schedule ad-hoc calls as needed.
If I bring on multiple fellows or external contractors, I'll add a weekly all-hands to make sure everyone has situational awareness.
Essential:
Preferred:
-Specific chokepoint experience: Familiarity with cloud labs, CROs, live pathogen repositories, cell-free systems, plasmid synthesis, or lab equipment ecosystems is a plus.
-Technical writing and translation skills: You can synthesize complex technical findings for policymakers and other non-technical stakeholders.
While fellows will have some flexibility in selecting the chokepoint they focus on (based on their background and interests), we've designed specific projects and open research questions for each chokepoint to shape their contributions.
Implementing SL4/5 and searching for differentially defense-favored security tools.
I love asynchronous collaboration and I'm happy to provide frequent small directional feedback, or do thorough reviews of your work with a bit more lead time. A typical week should look like either trying out a new angle on a problem, or making meaningful progress towards productionizing an existing approach.
Essential:
Preferred:
Mentor(s) will talk through project ideas with scholar, or scholar will pick from a list of projects.
GDM stream focused on scheming risk, AI control, monitoring, monitorability, and loss-of-control evaluations. Probably running in-person in London.
I'm generally quite hands-off. I propose projects that I think matter for AGI safety and are tractable, to set scholars up for success. I then expect scholars to fully own the project, and update / consult me as needed.
By default we'd meet once a week to discuss the project for 30 min - 1 hour. I see my role as giving feedback on the project direction, stress-testing or advising on design / prioritisation decisions, and occasionally suggesting experiments or methodological improvements (which you should treat as suggestions / advice, not orders!).
You can also book ad-hoc meetings with me, ping me on Slack, or send me docs / paper drafts for review.
I also offer scholars to meet with me once a month for 30 min to discuss career stuff, skill-building, feedback on their progress, or anything else.
Preferred technical skills:
I'll propose ~5 projects for scholars to choose from. I am also open to scholar-proposed projects if they are well articulated, promising, and align with my research interests.
Computational/modelling problems in biosecurity.
typically 1 hour weekly meetings by default. I typically respond on slack quite quickly - some weeks that I am not available. You are welcome to chat to my phd students too!
Computational experience e.g. Python
OR statistical modelling
interest in biosecurity
We will construct a project together that best suits the skills and interests of the fellow and what I can reasonably be helpful for.
MATS Research phase provides scholars with a community of peers.

Scholars work out of a shared office and are supported by the Community Team.
MATS alumni report that the connections with peers that they made during MATS have had the largest impact on them years later. Our full-time Community Team works to facilitate these connections and also provide general well-being support. Weekly lightning talks, scholar-led discussion groups, game nights, and outings to SF are some examples of MATS events.