Position Details
About this role
This role involves developing and optimizing large-scale ML systems, focusing on large language models and inference performance within AWS infrastructure.
Key Responsibilities
- Develop ML systems end-to-end
- Optimize inference performance
- Build reliable ML infrastructure
- Collaborate with scientists and engineers
- Manage GPU and accelerator hardware
Technical Overview
The environment includes AWS cloud infrastructure, GPU hardware, custom accelerators, and distributed ML systems, emphasizing inference optimization and scalable deployment.
Ideal Candidate
The ideal candidate is a mid-level machine learning engineer with experience in large language models, inference optimization, and distributed systems. They are proficient with GPU hardware and cloud environments, especially AWS.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Keywords for Your Resume
Deal Breakers
No experience with large language models, Lack of inference optimization skills, No background in distributed systems
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile