Anthropic is pioneering the development of safe, ethical, and powerful artificial intelligence systems. As a Research Engineer on the Pretraining team, you'll be at the forefront of developing next-generation large language models. This role uniquely combines cutting-edge research with practical engineering challenges.
The position offers an opportunity to work on groundbreaking AI technology while prioritizing safety and ethical considerations. You'll be part of a collaborative team conducting research on model architecture, algorithms, and data processing, while also optimizing training infrastructure and developing essential tools.
Key projects include optimizing novel attention mechanisms, comparing Transformer variants' compute efficiency, preparing large-scale datasets, scaling distributed training to thousands of GPUs, and creating interactive visualizations of model internals.
Anthropic operates as a cohesive team focused on large-scale research efforts, viewing AI research as an empirical science. The company offers competitive compensation, benefits including equity donation matching, generous vacation and parental leave, and a collaborative office space in San Francisco.
The ideal candidate will have advanced degrees in Computer Science or related fields, strong software engineering skills, and expertise in Python and deep learning frameworks. You should be passionate about AI safety, comfortable balancing research with engineering constraints, and eager to contribute to a team working on transformative AI systems.
Anthropic values diversity and strongly encourages applications from candidates of all backgrounds, including underrepresented groups in tech. The company sponsors visas and offers a hybrid work environment requiring at least 25% office presence.
This role is perfect for someone who views research and engineering as complementary disciplines, has ambitious goals for AI safety, and is committed to creating the best long-term outcomes in AI development.