MATS Summer 2026 applications are now open!
<https://matsprogram.org/apply?utm_source=corpora&utm_medium=&utm_campaign=s26>
We believe reducing risks from unaligned AI is one of the world's most
urgent and talent-constrained challenges, and that ambitious people from a
wide range of backgrounds can meaningfully contribute to this work. That's
why we're training the next generation of AI alignment, interpretability,
security, and governance researchers.

Running June-August 2026, MATS’ 12-week research program provides
everything you need to launch your career in AI safety. This includes
field-leading research mentorship, funding ($15k stipend + $12k compute),
Berkeley/London offices (depending on mentor preference), housing,
talks/workshops with AI experts, and a global network of peers.

MATS has accelerated 450+ researchers so far. Among alumni who graduated
before 2025, 80% are working directly in AI safety/security and 10% have
co-founded active AI safety startups. Participants have coauthored 150+
papers, with 7,800+ citations
<https://scholar.google.com/citations?user=VgJaUK4AAAAJ&hl=en>, and rate
our program 9.4/10. Our mentors include world-class researchers from
Anthropic, Google DeepMind, OpenAI, UK AISI, GovAI, Redwood, METR, Apollo
Research, Goodfire, RAND, AI Futures Project, and more.

*Apply by January 18 AoE to be considered!* Visit our website
<https://matsprogram.org/apply?utm_source=corpora&utm_medium=&utm_campaign=s26>
for details.


Help us spread the word with people you know who'd be a good fit, and feel
free to reach out if you have any questions!

Best,
Eric

--

Eric Dhan
Operations Generalist, MATS
LinkedIn <https://www.linkedin.com/in/eric-dhan/>
_______________________________________________
Corpora mailing list -- [email protected]
https://list.elra.info/mailman3/postorius/lists/corpora.list.elra.info/
To unsubscribe send an email to [email protected]

Reply via email to