Anthropic is seeking a Research Scientist/Engineer to join their Finetuning Alignment team, focusing on developing AI systems that are reliable, truthful, and aligned with human values. This role combines cutting-edge AI research with practical engineering to minimize hallucinations and enhance truthfulness in language models.
The position offers an opportunity to work on significant challenges in AI safety and ethics, developing novel techniques for model truthfulness and accuracy. You'll be part of a collaborative team that approaches AI research as an empirical science, similar to physics and biology. The role involves creating sophisticated data curation pipelines, developing evaluation frameworks, and implementing retrieval-augmented generation systems.
Anthropic offers a competitive compensation package ranging from $280,000 to $425,000 USD, along with comprehensive benefits including flexible working hours and parental leave. The company maintains a hybrid work environment in San Francisco, requiring at least 25% office presence.
The ideal candidate will have advanced degrees in Computer Science or ML, strong Python skills, and experience with language model finetuning. They should be passionate about AI safety and have a track record in developing systems for model accuracy and truthfulness. The role presents an excellent opportunity to contribute to ensuring AI systems remain reliable and ethical while advancing the field of AI safety.
Working at Anthropic means joining a cohesive team focused on large-scale research efforts rather than smaller, specific puzzles. The company values impact and collaboration, with frequent research discussions and a strong emphasis on communication skills. This is a chance to work on meaningful problems that could shape the future of AI development.