✦ Luna Orbit — AI & Machine Learning

Staff ML Engineer, Inference Platform

at General Motors

📍 Sunnyvale, California, United States of America Hybrid 💰 $185K – $270K USD / year Posted March 14, 2026
Salary $185K – $270K USD / year
Type Full-Time
Experience senior
Exp. Years 8+ years
Education Not specified
Category AI & Machine Learning

This role involves developing and scaling AI inference platforms for autonomous vehicle applications, focusing on high-performance distributed systems and GPU utilization.

  • Design core ML platform components
  • Optimize model serving
  • Lead large-scale AI infrastructure projects
  • Collaborate with ML engineers
  • Research state-of-the-art model deployment techniques

Requires expertise in machine learning systems, distributed computing, GPU hardware, cloud infrastructure, and container orchestration tools like Kubernetes and Docker.

The ideal candidate is a senior AI/ML engineer with over 8 years of experience in building scalable distributed machine learning systems, proficient in GPU hardware, cloud infrastructure, and container orchestration tools like Kubernetes and Docker.

8+ years of industry experienceExpertise in machine learning systems or backend servicesExperience with distributed systemsKnowledge of GPU hardware
Open source contributionsAI infrastructureModel optimizationScalabilityReal-time inference
KubernetesDockerGPU hardware (H100A100B200)
Machine LearningDistributed SystemsModel ServingBackend SoftwareGPU UtilizationCloud ComputingKubernetesDocker
Machine LearningDistributed SystemsModel ServingBackend SoftwareGPU UtilizationCloud ComputingHigh Performance ComputingPythonC++KubernetesDocker
Problem-solvingTechnical LeadershipCollaborationInnovationCommunication
Industry Automotive
Job Function Building and scaling AI inference platforms for autonomous vehicles
Role Subtype AI & Machine Learning
Tech Domains Machine Learning, Distributed Systems, Cloud Computing, High Performance Computing
Clearance Required None
Visa Sponsorship No
ML EngineerMachine LearningDistributed SystemsModel ServingBackend SoftwareGPU UtilizationCloud ComputingKubernetesDockerAI InfrastructureModel OptimizationReal-time inferenceHigh Performance ComputingOpen sourceSunnyvaleAImachine learningdistributed systemsmodel servingGPU utilizationcloud computingAI infrastructurereal-time inferencehigh performance computing

Less than 8 years of experience, Lack of expertise in distributed systems or GPU hardware, No experience with Kubernetes or Docker, Inability to work in hybrid environment

Apply for this Position →

Get matched to jobs like this

Luna finds roles that fit your skills and career goals — no endless scrolling required.

Create a Free Profile