Anthropic is at the forefront of developing safe and beneficial AI systems, with a mission focused on creating reliable, interpretable, and steerable AI. As a Trust and Safety Machine Learning Engineer, you'll play a crucial role in building safety and oversight mechanisms for AI systems. The position combines technical ML expertise with a strong focus on ethical considerations and user well-being.
The role involves developing and implementing machine learning models for detecting harmful behaviors and ensuring compliance with terms of service. You'll work with cutting-edge AI technology while focusing on trust and safety applications, including behavioral classifiers and anomaly detection systems. The position requires both technical excellence and strong communication skills to bridge technical and non-technical stakeholders.
Anthropic operates as a public benefit corporation, emphasizing the importance of AI safety and ethical considerations in their work. The company approaches AI research as an empirical science, similar to physics and biology, and values collaborative work on large-scale research efforts over smaller, isolated projects. Their research portfolio includes significant work on GPT-3, Circuit-Based Interpretability, Multimodal Neurons, and other cutting-edge AI technologies.
The company offers a competitive compensation package starting from $340,000 to $425,000 USD, along with comprehensive benefits including visa sponsorship, flexible working arrangements, and generous leave policies. The hybrid work environment requires at least 25% office presence in San Francisco, fostering a collaborative atmosphere while maintaining flexibility. This is an excellent opportunity for experienced ML engineers who want to make a meaningful impact on the safe development of AI technology.