✦ Luna Orbit — Software Engineering

Software Engineer, Inference Deployment

at Anthropic

📍 San Francisco, CA | New York City, NY | Seattle, WA Hybrid Posted March 07, 2026
Type Not Specified
Experience mid
Exp. Years 5+ years
Education Not specified
Category Software Engineering

This role involves designing and building scalable deployment infrastructure for AI inference models across hardware accelerators, ensuring continuous, unattended deployment with high efficiency.

  • Own deployment orchestration
  • Improve deployment scheduling
  • Extend deployment observability
  • Drive down cycle time
  • Optimize fleet rollout strategies

Focuses on AI inference deployment systems, resource management, automation, and monitoring across GPU, TPU, and Trainium hardware, with a strong emphasis on systems engineering and scalability.

The ideal candidate is a software engineer with over 5 years of experience in building deployment and automation systems at scale, particularly involving inference deployment on accelerators like GPUs, TPUs, and Trainium. They should have strong systems design skills and experience managing resource-constrained environments.

5+ years of experience building deploymentreleaseor delivery infrastructure at scaleStrong software engineering skillsExperience designing systems managing complex state machines and multi-stage pipelinesExperience with deployment systems where resource constraints shape the design
Experience with GPUsTPUsTrainiumKnowledge of continuous deploymentMonitoring and observability toolsAutomation tooling
GPUTPUTrainiumDashboardsPipeline architectures
Deployment systemsInference deploymentGPUTPUTrainiumResource managementAutomationPipeline architecturesOrchestrationMonitoring dashboardsSoftware engineering
Deployment systemsInference deploymentGPUTPUTrainiumResource managementAutomationPipeline architecturesOrchestrationDeployment infrastructureMonitoring dashboardsSoftware engineering
Problem-solvingAutomationCollaborationResource managementCommunication
Industry Technology
Job Function Build and manage AI inference deployment infrastructure at scale
Deployment systemsInference deploymentGPUTPUTrainiumResource managementAutomationPipeline architecturesOrchestrationMonitoring dashboardsSoftware engineeringValidationScaling infrastructureContinuous deploymentResource constraints

Less than 5 years of relevant experience, Lack of experience with deployment systems at scale, No experience with GPU, TPU, or Trainium hardware

Apply for this Position →

Get matched to jobs like this

Luna finds roles that fit your skills and career goals — no endless scrolling required.

Create a Free Profile