AWS Neuron is seeking a Machine Learning Compiler Engineer to join their innovative team working on the SDK that optimizes ML models for AWS Inferentia and Trainium custom chips. This role offers an exciting opportunity to be at the forefront of AI revolution, working on transforming complex ML models from frameworks like PyTorch, TensorFlow, and JAX for deployment on AWS's specialized hardware.
As a Compiler Engineer, you'll tackle challenging optimization problems for various ML model families, including large language models like Llama and Deepseek, stable diffusion, and vision transformers. The role requires deep technical understanding of ML models and compiler optimization techniques to generate optimal implementations.
AWS offers a collaborative environment with a diverse team that values knowledge-sharing and mentorship. You'll work alongside chip architects, runtime engineers, and ML teams, contributing to both proprietary and open-source software projects. The position provides competitive compensation ($129,300-$223,600 based on location) and comprehensive benefits.
The ideal candidate brings 3+ years of software development experience, strong object-oriented programming skills, and preferably experience with compiler design or ML frameworks. You'll be part of AWS's mission to democratize AI access while working on cutting-edge technology that impacts developers worldwide.
This role combines technical depth with business impact, offering opportunities to work with the latest ML technologies while solving complex engineering challenges. Join AWS to shape the future of AI infrastructure and be part of a team that values innovation, diversity, and professional growth.