Position Details
About this role
This role involves developing and scaling AI inference platforms supporting autonomous vehicle AI workloads, focusing on model serving, distributed systems, and GPU optimization.
Key Responsibilities
- Build inference platform
- Implement model serving strategies
- Optimize GPU utilization
- Develop monitoring and observability
- Collaborate with ML teams
Technical Overview
The position requires expertise in Python, C++, distributed systems, GPU utilization, and model serving frameworks like Triton and RayServe, with a focus on scalable, cloud-agnostic AI infrastructure.
Ideal Candidate
The ideal candidate is a senior ML infrastructure engineer with over 5 years of experience in building scalable, cloud-agnostic inference platforms, proficient in Python and C++, with hands-on experience in model serving frameworks like Triton and RayServe. They excel in distributed systems and GPU optimization.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Keywords for Your Resume
Deal Breakers
Less than 5 years experience in ML systems, No experience with model serving frameworks, Lack of proficiency in Python or C++, No knowledge of distributed systems
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile