Position Details
About this role
This role involves designing and optimizing container and cloud infrastructure for NVIDIA's inference microservices, focusing on GPU workloads, multi-architecture builds, and scalable deployment strategies.
Key Responsibilities
- Design and implement container strategies
- Optimize container performance
- Manage multi-arch builds
- Deploy GPU workloads in Kubernetes
- Collaborate with research and backend teams
Technical Overview
The technical environment includes Docker, Kubernetes, Helm, CUDA, and NVIDIA device plugins, aimed at building high-performance, scalable containerized AI inference systems.
Ideal Candidate
The ideal candidate is a senior software engineer with expertise in containerization, cloud infrastructure, and GPU workloads, proficient in Python and Kubernetes, with experience in performance tuning and multi-cluster deployments.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Keywords for Your Resume
Deal Breakers
Lack of experience with containers or Kubernetes, No GPU workload experience, Insufficient Python skills, No background in container performance tuning, Lack of collaboration skills
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile