AWS Neuron is seeking a talented Software Engineer to join their Machine Learning Applications (ML Apps) team, focusing on the complete software stack for AWS Inferentia and Trainium cloud-scale machine learning accelerators. This role presents an exciting opportunity to work at the forefront of machine learning infrastructure, specifically with massive-scale language models like LLama3, Mixtral, and DBRX.
The position involves close collaboration with chip architects, compiler engineers, and runtime engineers to develop and optimize distributed training solutions. You'll be responsible for implementing distributed training support in frameworks like PyTorch and JAX, while ensuring maximum performance on AWS Trainium and Inferentia silicon.
The ideal candidate will bring strong software development skills combined with deep ML knowledge. You'll be working in an inclusive environment that values diversity and work-life balance. Amazon offers comprehensive benefits, including medical and financial packages, and emphasizes career growth through mentorship and knowledge sharing.
The role offers competitive compensation ranging from $129,300 to $223,600 based on location and experience. You'll be part of a team that embraces Amazon's 16 Leadership Principles, including seeking diverse perspectives and earning trust. The position provides opportunities to work on cutting-edge ML infrastructure while maintaining a healthy work-life balance.
Join a team that's dedicated to supporting new members and fostering an environment of continuous learning and professional development. You'll have the chance to shape the future of machine learning infrastructure while working with some of the most advanced AI accelerator technologies in the industry.