Research Engineer / Scientist, Alignment Science jobs in United States
cer-icon
Apply on Employer Site
company-logo

Anthropic · 2 weeks ago

Research Engineer / Scientist, Alignment Science

Anthropic is a public benefit corporation dedicated to creating reliable and beneficial AI systems. The Research Engineer on Alignment Science will conduct exploratory research on AI safety, focusing on the risks from powerful future systems, and collaborate with various teams to develop techniques for ensuring the safety and alignment of AI models.

Artificial Intelligence (AI)Foundational AIGenerative AIInformation TechnologyMachine Learning
check
H1B Sponsorednote

Responsibilities

Build and run elegant and thorough machine learning experiments to help us understand and steer the behavior of powerful AI systems
Contribute to exploratory experimental research on AI safety, with a focus on risks from powerful future systems
Develop techniques to keep highly capable models helpful and honest, even as they surpass human-level intelligence in various domains
Create methods to ensure advanced AI systems remain safe and harmless in unfamiliar or adversarial scenarios
Create model organisms of misalignment to improve our empirical understanding of how alignment failures might arise
Build and align a system that can speed up & improve alignment research
Understand and document the highest-stakes and most concerning emerging properties of models through pre-deployment alignment and welfare assessments
Develop robust defenses against adversarial attacks, comprehensive evaluation frameworks for model safety, and automated systems to detect and mitigate potential risks before deployment
Investigate and address potential model welfare, moral status, and related questions
Test the robustness of our safety techniques by training language models to subvert our safety techniques, and seeing how effective they are at subverting our interventions
Run multi-agent reinforcement learning experiments to test out techniques like AI Debate
Build tooling to efficiently evaluate the effectiveness of novel LLM-generated jailbreaks
Write scripts and prompts to efficiently produce evaluation questions to test models’ reasoning abilities in safety-relevant contexts
Contribute ideas, figures, and writing to research papers, blog posts, and talks
Run experiments that feed into key AI safety efforts at Anthropic, like the design and implementation of our Responsible Scaling Policy

Qualification

Machine LearningAI Safety ResearchReinforcement LearningPythonResearch EngineeringEmpirical ResearchCollaborationCommunication SkillsTechnical WritingProblem Solving

Required

Have significant software, ML, or research engineering experience
Have some experience contributing to empirical AI research projects
Have some familiarity with technical AI safety research
Prefer fast-moving collaborative projects to extensive solo efforts
Pick up slack, even if it goes outside your job description
Care about the impacts of AI
Education requirements: We require at least a Bachelor's degree in a related field or equivalent experience

Preferred

Have experience authoring research papers in machine learning, NLP, or AI safety
Have experience with LLMs
Have experience with reinforcement learning
Have experience with Kubernetes clusters and complex shared codebases
Candidates Need Not Have 100% of the skills needed to perform the job
Formal certifications or education credentials

Benefits

Equity
Benefits
Incentive compensation
Optional equity donation matching
Generous vacation and parental leave
Flexible working hours

Company

Anthropic

twittertwittertwitter
company-logo
Anthropic is an AI research company that focuses on the safety and alignment of AI systems with human values.

H1B Sponsorship

Anthropic has a track record of offering H1B sponsorships. Please note that this does not guarantee sponsorship for this specific role. Below presents additional info for your reference. (Data Powered by US Department of Labor)
Distribution of Different Job Fields Receiving Sponsorship
Represents job field similar to this job
Trends of Total Sponsorships
2025 (105)
2024 (13)
2023 (3)
2022 (4)
2021 (1)

Funding

Current Stage
Late Stage
Total Funding
$33.74B
Key Investors
Lightspeed Venture PartnersGoogleAmazon
2025-09-02Series F· $13B
2025-05-16Debt Financing· $2.5B
2025-03-03Series E· $3.5B

Leadership Team

leader-logo
Dario Amodei
CEO & Co-Founder
linkedin
leader-logo
Daniela Amodei
President and co-founder
linkedin
Company data provided by crunchbase