Position Details
About this role
This role focuses on building and optimizing distributed systems for AI inference, including request routing, autoscaling, and multi-region deployment across cloud platforms.
Key Responsibilities
- Design request routing algorithms
- Autoscale compute fleet
- Build deployment pipelines
- Integrate new AI accelerators
- Optimize inference performance
Technical Overview
Involves distributed systems engineering, Kubernetes, cloud infrastructure, ML inference optimization, and deployment pipelines.
Ideal Candidate
The ideal candidate is a results-oriented software engineer with extensive experience in distributed systems, Kubernetes, and large-scale ML inference deployment. They are flexible, impact-driven, and comfortable working across cloud platforms.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Keywords for Your Resume
Deal Breakers
Lack of experience with distributed systems or Kubernetes, No background in ML inference or deployment pipelines, Inability to work in fast-paced, impact-driven environments
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile