Research Scientist/Engineer - Alignment Finetuning

AI research company focused on creating reliable, interpretable, and steerable AI systems for safe and beneficial use.
$280,000 - $425,000
Machine Learning
Senior Software Engineer
Hybrid
101 - 500 Employees
5+ years of experience
AI

Description For Research Scientist/Engineer - Alignment Finetuning

Anthropic is seeking a Research Scientist/Engineer to join their Alignment Finetuning team, focusing on developing AI systems that are safe and beneficial for society. This role combines cutting-edge ML research with practical implementation, working on training language models that demonstrate better moral reasoning, improved honesty, and good character.

The position offers an opportunity to work with a collaborative team of researchers, engineers, and policy experts in a company that approaches AI research as an empirical science. You'll be responsible for developing novel finetuning techniques, implementing advanced training pipelines, and creating evaluation frameworks to measure and improve model alignment properties.

The ideal candidate will have advanced education in Computer Science or ML, strong programming skills (particularly in Python), and experience with ML model training and experimentation. The role requires both technical expertise and the ability to collaborate effectively across teams.

Anthropic offers competitive compensation ($280,000 - $425,000), flexible working arrangements, and a strong benefits package. The company values diversity and encourages applications from candidates with varied backgrounds and perspectives. As a public benefit corporation, Anthropic is committed to developing AI systems that are reliable, interpretable, and beneficial for society.

Working at Anthropic means joining a cohesive team focused on large-scale research efforts rather than smaller, specific puzzles. The company's research builds on significant work in areas like GPT-3, Circuit-Based Interpretability, and AI Safety, making this an excellent opportunity for those interested in advancing the field of beneficial AI.

Last updated 4 days ago

Responsibilities For Research Scientist/Engineer - Alignment Finetuning

  • Develop and implement novel finetuning techniques using synthetic data generation and advanced training pipelines
  • Train models to have better alignment properties including honesty, character, and harmlessness
  • Create and maintain evaluation frameworks to measure alignment properties in models
  • Collaborate across teams to integrate alignment improvements into production models
  • Develop processes to help automate and scale the work of the team

Requirements For Research Scientist/Engineer - Alignment Finetuning

Python
  • MS/PhD in Computer Science, ML, or related field, or equivalent experience
  • Strong programming skills, especially in Python
  • Experience with ML model training and experimentation
  • Track record of implementing ML research
  • Strong analytical skills for interpreting experimental results
  • Experience with ML metrics and evaluation frameworks
  • Excellence at turning research ideas into working code
  • Ability to identify and resolve practical implementation challenges

Benefits For Research Scientist/Engineer - Alignment Finetuning

Visa Sponsorship
Equity
  • Competitive compensation and benefits
  • Optional equity donation matching
  • Generous vacation and parental leave
  • Flexible working hours
  • Office space for collaboration

Interested in this job?

Jobs Related To Anthropic Research Scientist/Engineer - Alignment Finetuning

Software Engineer

Senior Software Engineer role at Anthropic focusing on building large-scale ML systems with emphasis on safety and reliability.

Biosecurity Research Engineer

Senior Machine Learning Engineer role focused on AI safety and biosecurity research at Anthropic

Safeguards Research Engineer

Senior AI Safety Research Engineer position at Anthropic, focusing on developing and implementing safety measures for advanced AI systems.

Research Engineer, Frontier Red Team

Senior Research Engineer position at Anthropic focusing on AI safety evaluation and implementation of responsible scaling policies for frontier AI models.

Research Engineer, Frontier Red Team

Senior Research Engineer position at Anthropic focusing on AI safety evaluation and risk assessment for frontier AI models.