The MATS Biosecurity Track supports research at the intersection of advanced AI and catastrophic biological risk. We are launching this track because the threat model has shifted: biological foundation models, LLMs with growing wet-lab uplift, and AI-accelerated design tools are compressing timelines on capabilities the existing biosecurity stack was not built to absorb. We want fellows pursuing technical work that has a realistic chance of meaningfully shifting outcomes within the next 6–12 months.
The track spans six research areas. Fellows are matched to mentors based on fit, and projects are scoped to produce concrete artifacts – papers, evals, prototypes, or policy analyses – by program end.
Metagenomic surveillance pipelines for pandemic-grade pathogen detection; genomic language models for novelty detection and improved signal/noise at the front end of the surveillance stack.
AI-accelerated discovery of antiviral peptides under pandemic-response constraints, paired with realistic analysis of the manufacturing and supply-chain bottlenecks that determine whether candidates actually reach patients.
Function-based DNA sequence screening using mechanistic interpretability and ML on biological foundation models — classifiers that catch hazardous sequences, including engineered and AI-designed variants meant to evade homology-based screens.
Engineering work on emergency biodefense infrastructure: PPE, filtration, far-UVC, decontamination, and improvised protective systems for worst-case scenarios. Deliverables here are often physical or quasi-physical.
Policy and forecasting work on AI-bio: evaluating policy levers, forecasting when AI trivializes specific offensive or defensive capabilities, and analyzing deterrence via physical chokepoints (synthesis screening governance, cloud-lab access controls).
Red-teaming biological AI models for dangerous capabilities, building technical defenses (genetic engineering attribution, data governance), and developing dangerous-capability evaluations for frontier bio-AI.
We expect fellows to engage seriously with infohazard considerations and to operate within a publication and disclosure framework we'll work through together early in the program. If you're uncertain whether your background fits, apply anyway and tell us how you think about the threat model — that reasoning is more informative to us than credentials.
We anticipate that strong candidates will come from a variety of backgrounds, including biology, AI safety, public health, epidemiology, machine learning, engineering, chemistry, biosafety, biosecurity, and national security. If you're uncertain whether your background fits, apply anyway and tell us how you think about the threat model; that reasoning is more informative to us than credentials.
This stream is primarily focused on research into physical defenses against engineered pathogens, aiming to inform decisions about PPE stockpiling and distribution approaches, improve improvised PPE and bioshelter scale-up, and reach rapid conclusions on how much to prioritize other areas of physical biodefense (agriculture, emergency response, etc.). We are also open to strategic research into the use of bioweapons by AI or AI-human teams as part of takeover strategies and how this might inform preparedness.
One 30-60 min weekly meeting by default. We’re active on slack and can usually respond to quick questions there within the work day. For more substantive async engagement, especially project feedback, google doc comments are probably best.
The most important attribute is being generative when attacking a problem and willing to try a bunch of angles—reaching out to experts, contacting companies, prototyping stuff on your own, etc. While you’d develop a research output, we expect the fellows best suited to this workstream will adopt a dogged attitude, keeping an eye out for opportunities to apply findings to future biosecurity projects like starting a new org or contributing to work in an existing org.
Fellows should also be familiar with developing BOTECs, willing to quickly get up to speed in new technical areas, capable of working independently, and happy to pivot in response to findings or feedback.
A background in the physical sciences or engineering may be helpful, but is definitely not a requirement.
We’ll provide fellows with a short list of projects and will meet with them to discuss which they feel most excited about / best suited for.
I'm mentoring projects that apply AI advances to core biosecurity challenges — early detection and attribution of biological threats, characterizing AI-enabled bioweapons uplift, accelerating medical countermeasure design, and building biosecurity-by-design into frontier biological AI tools. Fellows have wide latitude to scope their own project within (or adjacent to) these themes, including ideas I haven't yet considered, and are expected to drive the technical work independently. My comparative advantage is high-level strategic direction grounded in biosecurity, pandemic preparedness, epidemiology, and US R&D policy rather than hands-on ML or software engineering guidance.
I anticipate meeting with the fellow for 30 minutes each Monday and Thursday for high-level guidance. We can coordinate on Slack or email, and I will typically respond within a day for quick questions or conceptual feedback. I will not debug code. Expect asynchronous iteration on experiment design and results between meetings. Scholars can also schedule ad-hoc calls if they're stuck or want to brainstorm.
Essential skills:
Preferred skills:
Not a good fit:
While I am hopeful that my partner fellow will pick a project in one of the areas I've delineated above, fellows have wide latitude to decide on a specific project to pursue, subject to my agreement (I want to make sure they choose a path that is both useful and viable). I am also very open to new ideas from fellows that I have not already thought of-particularly ones that address some of the challenges raised above. If needed, my assigned fellow can experiment with exploring multiple projects during the first week before picking one to focus on.
This mentor also has a stream in the Strategy and Forecasting track
This stream focuses on how advanced AI could enable new and dangerous bio technologies, and on assessing when risks become tractable or urgent as those capabilities arrive.
Half-hour one-on-one weekly meetings by default, with the option to extend or add ad-hoc calls when useful. I'm active on Slack and typically respond within a day for quick questions. I'm happy to read drafts and leave written feedback async between meetings.
Essential:
Preferred:
I'll talk with the fellow about what they're interested in, and we'll pick a broad area together from a few directions I'd want to pitch. From there we'll work together to scope something sharp and well-defined, with me leaning on my sense of what's tractable and high-value. The fellow then runs with the project, and we adjust as it develops.
The stream focuses on evaluating and/or mitigating catastrophic risk emerging from dangerous scientific capabilities in frontier AI systems, with an emphasis on the challenges that emerge from lab integrations and novel science. Potential research directions include evaluation design, risk mitigations and evaluation science.
We can schedule a weekly 1h meeting, for general progress updates, share result and overall guidance. I would be reachable on Slack as well for async comms. Happy to jump on ad-hoc calls for specific discussions or pair coding/debugging. I am based in London and I work UK hours (10am-7pm), but I also visit the US (Boston) a few times a year.
Essential
Preferred
Not a good fit:
I will work with the fellow to find the right project that suits their interest within the directions spelled out above. I will pitch a few project ideas and support the fellow in making the decision. I also welcome project suggestions; in those cases I would work with the fellow to scope it appropriately.
I’m interested in two broad projects focused on improving current detection efforts at SecureBio. The first is to characterize when AI-bio or general AI tools are actually useful for large-scale metagenomic detection, including tradeoffs between compute cost, sequencing cost, model type, model size, and pipeline stage. The second is to explore genomic language models as novelty detectors—for example, using perplexity-style metrics to flag surprising sequences—and to evaluate whether this approach can complement traditional bioinformatics systems in a cost-effective, sensitive, and interpretable way.
By default, we'll mostly collaborate via a standing weekly meeting (~1 hour), wherein we'll discuss recent progress and next directions. I'm available via Slack for quick back-and-forth on ideas, sanity checks, and unblocking (data access, etc.), but will rely on the fellow to manage their own implementations, code review, debugging, etc.
Essential:
Preferred:
Not a good fit:
I'll determine which of the two broad project ideas we're running with based on SecureBio Detection needs, which fellows match to me, etc. Within that broad project, I'll guide with what I think is helpful / interesting / relevant to SecureBio Detection, and I expect the fellow to have both autonomy and responsibility to pick concrete work directions.
Fourth Eon is developing adaptive, AI-native safeguards across the biotechnology stack, with a focus on function-based DNA synthesis screening. Fellows in this stream will work on technical research projects at the intersection of AI and biosecurity. Projects span topics like mechanistic interpretability of protein foundation models, bio model evaluations for biosecurity-relevant capabilities, and agentic sequence analysis workflows.
I typically schedule a standing weekly 1:1 meeting with each fellow, and also hold a weekly research group meeting. Beyond that I am available on Slack and can find additional time for calls outside of scheduled meetings.
Note that as part of our Safe and Responsible Research Framework we require fellows to sign a fellowship agreement covering confidentiality and pre-publication review for dual-use risks. This is common practice in biosecurity research and allows us to work freely together on sensitive material.
Required:
• Prior technical research experience
• Strong critical thinking and creative problem-solving abilities
• The integrity and judgment to responsibly carry out sensitive research
• A good understanding of the basics of biomolecular sequence, structure, and function
• Expertise in one or more of the following domains:
bioinformatics, computational biology, structural biology, biochemistry, molecular biophysics, protein engineering, biosecurity, AI/ML, or a related field
• Proficiency with Python
Preferred:
• Hands-on experience with testing biological AI models
• Have built model evaluations / benchmarks
• Experience with mechanistic interpretability techniques
• Biosecurity context awareness
Fellows who are interested in our research area should think of potential project ideas that leverage their strengths and interests. I will work individual fellows to identify a specific project that matches their background and interests and is aligned with our overall research direction, and to refine the scope and objectives of the project.
This stream will focus on projects related to biosecurity countermeasures.
1 hour weekly meetings by default for high-level guidance. Onboarding to our slack, which has access to the entire Blueprint Biosecurity team. Can be reached async every day and can meet as needed.
Successful fellows likely have some sort of a technical background (e.g. technical undergraduate degree), have familiarity with reading research papers, are agentic, prioritize truthseeking, and focused on maximizing impact.
Relevant biosecurity experience is a plus, but not required.
We are very focused on advancing priority countermeasures quickly. We anticipate having projects only focus on these areas:
Within these areas there is some latitude for different projects.
This stream focuses on lead independent research in one of six chokepoints for biotech governance: live pathogen repositories, CROs, cloud labs, cell-free expression systems, plasmid vendors, or secondhand lab equipment.
On high-conviction areas, you'll tackle specific open research questions and assess interventions; on low-conviction areas, you'll conduct deep dives to determine whether they're worth pursuing. Your findings will directly shape Sentinel's grantmaking strategy and provide strategic guidance to the broader biosecurity community.
I'll hold a one-hour weekly check-in by default, with higher frequency during onboarding. I'm available via Slack with quick turnarounds on async messages, and you can schedule ad-hoc calls as needed.
If I bring on multiple fellows or external contractors, I'll add a weekly all-hands to make sure everyone has situational awareness.
Essential:
Preferred:
-Specific chokepoint experience: Familiarity with cloud labs, CROs, live pathogen repositories, cell-free systems, plasmid synthesis, or lab equipment ecosystems is a plus.
-Technical writing and translation skills: You can synthesize complex technical findings for policymakers and other non-technical stakeholders.
While fellows will have some flexibility in selecting the chokepoint they focus on (based on their background and interests), we've designed specific projects and open research questions for each chokepoint to shape their contributions.
Computational/modelling problems in biosecurity.
typically 1 hour weekly meetings by default. I typically respond on slack quite quickly - some weeks that I am not available. You are welcome to chat to my phd students too!
Computational experience e.g. Python
OR statistical modelling
interest in biosecurity
We will construct a project together that best suits the skills and interests of the fellow and what I can reasonably be helpful for.
Therapeutics may have durable advantages over pathogens even in the limit of technological progress. How can therapeutic development and manufacturing be made resilient under biorisk scenarios? How can AI progress be maximally leveraged for defense?
I expect we will spend some time at the beginning scoping out a project that is a good fit for the fellow's background and interests. Then, project supervision will depend strongly on the nature of the project. Generally, I expect the fellow to take ownership of the work, with regular mentorship and feedback to maintain alignment and help resolve challenges as they arise.
Essential:
Preferred:
Not a good fit:
We will jointly define the exact project with the fellow, based on their background, interests, and comparative strengths, as well as our current priorities. I expect strong fellows may have their own questions and ideas, but we will provide substantial guidance early on to help turn those ideas into a clear, useful, and realistically scoped project.
In the first phase, we will discuss several possible directions, identify where the fellow can make the strongest contribution, and agree on concrete outputs. Once the project is scoped, I expect the fellow to take ownership of the work, with regular mentorship and feedback to keep it aligned and help overcome any difficulties.
This stream will work on projects that empirically assess national security threats of AI misuse (CBRN terrorism and cyberattacks) and improve dangerous capability evaluations. Threat modeling applicants should have a skeptical mindset, enjoy case study work, and be strong written communicators. Eval applicants should be able and excited to help demonstrate concepts like sandbagging elicitation gaps in an AI misuse context.
The MATS Program is a 10-week research fellowship designed to train and support emerging researchers working on AI alignment, transparency and security. Fellows collaborate with world-class mentors, receive dedicated research management support, and join a vibrant community in Berkeley focused on advancing safe and reliable AI. The program provides the structure, resources, and mentorship needed to produce impactful research and launch long-term careers in AI safety.
MATS mentors are leading researchers from a broad range of AI safety, alignment, governance, field-building and security domains. They include academics, industry researchers, and independent experts who guide scholars through research projects, provide feedback, and help shape each scholar’s growth as a researcher. The mentors represent expertise in areas such as:
Key dates
Application:
The main program will then run from September 28th to December 4th, with the extension phase for accepted fellows beginning in December.
MATS accepts applicants from diverse academic and professional backgrounds - from machine learning, mathematics, and computer science to policy, economics, physics, cognitive science, biology, and public health, as well as founders, operators, and field-builders without traditional research backgrounds. The primary requirements are strong motivation to contribute to AI safety and evidence of technical aptitude, research potential, or relevant operational experience. Prior AI safety experience is helpful but not required.
Applicants submit a general application, applying to various tracks (Empirical, Theory, Strategy & Forecasting, Policy & Governance, Systems Security, Biosecurity, Founding & Field-Building.
In stage 2, applicants apply to streams within those tracks as well as completing track specific evaluations.
After a centralized review period, applicants who are advanced will then undergo additional evaluations depending on the preferences of the streams they've applied to before doing final interviews and receiving offers.
For more information on how to get into MATS, please look at this page.