Position Details
About this role
This role involves optimizing AI inference workloads on Intel GPUs, developing graph compilation flows, and tuning performance-critical GPU kernels for modern AI models.
Key Responsibilities
- Optimize AI inference workloads
- Develop graph compilation flows
- Write GPU kernels
- Profile and benchmark AI workloads
- Collaborate on hardware/software co-optimization
Technical Overview
The position requires expertise in C++, GPU architecture, neural network inference, and compiler technologies like MLIR and LLVM, with a focus on performance optimization and hardware/software co-design.
Ideal Candidate
The ideal candidate is a mid-level AI software engineer with at least 4 years of experience in GPU programming, neural network inference, and performance optimization, proficient in C++ and familiar with compiler technologies like MLIR and LLVM.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Keywords for Your Resume
Deal Breakers
Less than 4 years experience, No experience with GPU kernels or architecture, Lack of C++ development skills, No neural network inference experience
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile