Software Engineer, Model Inference

AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity.
$200,000 - $370,000
Backend
Senior Software Engineer
In-Person
501 - 1,000 Employees
3+ years of experience
AI

Description For Software Engineer, Model Inference

OpenAI is seeking a Software Engineer for Model Inference to join their Applied AI Engineering team in San Francisco. This role is crucial for scaling up critical inference infrastructure that efficiently services customer requests for state-of-the-art AI models like GPT-4 and DALL-E.

Key responsibilities include:

  • Collaborating with ML researchers, engineers, and product managers to productionize latest technologies
  • Implementing new techniques, tools, and architecture to enhance model performance, latency, throughput, and efficiency
  • Developing tools for identifying bottlenecks and instability sources, then designing and implementing solutions
  • Optimizing code and Azure VM fleet to maximize hardware utilization

Ideal candidates should have:

  • Understanding of modern ML architectures and optimization for inference
  • End-to-end problem-solving skills
  • At least 3 years of professional software engineering experience
  • Expertise in HPC technologies (InfiniBand, MPI, CUDA)
  • Experience with production distributed systems
  • Self-direction and ability to identify important problems
  • Humble attitude and eagerness to help colleagues

OpenAI offers a competitive salary range of $200K – $370K and is committed to diversity, equal opportunity, and providing reasonable accommodations to applicants with disabilities.

Join OpenAI in shaping the future of AI technology and ensuring its benefits are widely shared.

Last updated 7 months ago

Responsibilities For Software Engineer, Model Inference

  • Work with ML researchers, engineers, and product managers to bring latest technologies into production
  • Introduce new techniques, tools, and architecture to improve model performance, latency, throughput, and efficiency
  • Build tools to identify bottlenecks and sources of instability, then design and implement solutions
  • Optimize code and Azure VM fleet to maximize hardware utilization

Requirements For Software Engineer, Model Inference

Python
  • Understanding of modern ML architectures and optimization for inference
  • At least 3 years of professional software engineering experience
  • Expertise in core HPC technologies: InfiniBand, MPI, CUDA
  • Experience architecting, observing, and debugging production distributed systems
  • Self-direction and ability to identify important problems to work on

Interested in this job?

Jobs Related To OpenAI Software Engineer, Model Inference

Research Infrastructure Engineer - Post-Training

Senior Research Infrastructure Engineer role at OpenAI, building and optimizing systems for ChatGPT's post-training phase, offering $310K-$460K plus equity and comprehensive benefits.

Software Engineer, Financial Engineering

Senior Software Engineer role at OpenAI focusing on building and architecting financial engineering systems for billing and monetization.

Software Engineer, Internal Applications - Enterprise

Senior Software Engineer role at OpenAI focusing on internal applications and enterprise infrastructure automation using modern cloud technologies and DevOps practices.

Software Engineer, Financial Engineering

Senior Software Engineer role at OpenAI focusing on building and architecting next-generation billing and monetization systems.

Software Engineer, Internal Applications - Enterprise

Senior Software Engineer role at OpenAI focusing on internal applications and enterprise infrastructure automation