Position Details
About this role
This role focuses on training large AI models efficiently across multiple GPUs, improving pipeline performance, and contributing to open source AI frameworks.
Key Responsibilities
- Train large models
- Optimize training pipelines
- Contribute to open source
- Collaborate across teams
- Stay updated with training algorithms
Technical Overview
The environment involves distributed training pipelines, ML frameworks like PyTorch, TensorFlow, JAX, GPU kernel optimization, and large-scale AI model training.
Ideal Candidate
The ideal candidate is a highly skilled ML engineer with advanced knowledge of distributed training of large models, proficient in frameworks like PyTorch, TensorFlow, or JAX, and experienced in GPU optimization and large-scale AI research.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Keywords for Your Resume
Deal Breakers
Lack of experience with distributed training pipelines, No knowledge of ML frameworks (PyTorch, TensorFlow, JAX), No experience with large models or GPU optimization, Unable to work in the specified location
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile