Anthropic Fellows Program — Reinforcement Learning
💰 $70,000 – $100,000/yr
Job Description
About Anthropic
Anthropic's mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.
Anthropic Fellows Program Overview
The Anthropic Fellows Program is designed to foster AI research and engineering talent. We provide funding and mentorship to promising technical talent—regardless of previous experience. This is a unique opportunity for researchers and engineers passionate about AI safety and security to work on impactful projects with world-class mentors.
Fellows will primarily use external infrastructure (e.g., open-source models, public APIs) to work on empirical projects aligned with Anthropic's research priorities, with the goal of producing a public output such as a paper submission. In one of our earlier cohorts, over 80% of fellows produced publishable papers, demonstrating the program's strong track record of research output.
Program Structure & Timeline
- Duration: 4 months of full-time research
- Next cohort start date: July 20, 2026
- Application deadline: April 26, 2026
- Rolling applications accepted for later cohorts
- Multiple cohorts run each year
What You'll Receive
- Direct mentorship from experienced Anthropic researchers
- Weekly stipend: $3,850 USD / £2,310 GBP / $4,300 CAD + country-specific benefits
- Dedicated research compute funding (~$15,000/month)
- Additional funding for research expenses and tools
- Access to shared workspace in Berkeley, California or London, UK
- Connection to the broader AI safety and security research community
Ideal Candidate Profile
We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every listed requirement. Research shows that people from underrepresented groups are more prone to experiencing imposter syndrome and doubting their candidacy—we urge you not to exclude yourself prematurely. If you're interested in this work and passionate about AI safety, submit an application.
Interview Process
The selection process includes an initial application and reference check, technical assessments and interviews, and a dedicated research discussion. This multi-stage approach ensures we identify candidates with both technical depth and research potential.
This posting is specific to the Reinforcement Learning workstream. See the main Anthropic Fellows posting for additional workstream options and program details.
💰 Compensation not publicly listed. Market estimate for similar roles: from $70K, varying by experience and location. Program provides weekly stipend plus research funding as detailed above.