AI Infra Engineer
New Yesterday
Perplexity is an AI-powered answer engine founded in December 2022 and growing rapidly as one of the world’s leading AI platforms. Perplexity has raised over $1B in venture investment from some of the world’s most visionary and successful leaders, including Elad Gil, Daniel Gross, Jeff Bezos, Accel, IVP, NEA, NVIDIA, Samsung, and many more. Our objective is to build accurate, trustworthy AI that powers decision-making for people and assistive AI wherever decisions are being made. Throughout human history, change and innovation have always been driven by curious people. Today, curious people use Perplexity to answer more than 780 million queries every month–a number that’s growing rapidly for one simple reason: everyone can be curious.
We are looking for an AI Infra engineer to join our growing team. We work with Kubernetes, Slurm, Python, C++, PyTorch, and primarily on AWS. As an AI Infrastructure Engineer, you will work in a hybrid SRE/Dev Engineering capacity, partnering closely with our Infrastructure and Research teams to build, deploy, and optimize our large-scale AI training and inference clusters.
Responsibilities Design, deploy, and maintain scalable Kubernetes clusters for AI model inference and training workloads
Manage and optimize Slurm-based HPC environments for distributed training of large language models
Develop robust APIs and orchestration systems for both training pipelines and inference services
Implement resource scheduling and job management systems across heterogeneous compute environments
Benchmark system performance, diagnose bottlenecks, and implement improvements across both training and inference infrastructure
Build monitoring, alerting, and observability solutions tailored to ML workloads running on Kubernetes and Slurm
Respond swiftly to system outages and collaborate across teams to maintain high uptime for critical training runs and inference services
Optimize cluster utilization and implement autoscaling strategies for dynamic workload demands
Qualifications Strong expertise in Kubernetes administration, including custom resource definitions, operators, and cluster management
Hands-on experience with Slurm workload management, including job scheduling, resource allocation, and cluster optimization
Experience with deploying and managing distributed training systems at scale
Deep understanding of container orchestration and distributed systems architecture
High level familiarity with LLM architecture and training processes (Multi-Head Attention, Multi/Grouped-Query, distributed training strategies)
Experience managing GPU clusters and optimizing compute resource utilization
Required Skills Expert-level Kubernetes administration and YAML configuration management
Proficiency with Slurm job scheduling, resource management, and cluster configuration
Python and C++ programming with focus on systems and infrastructure automation
Hands-on experience with ML frameworks such as PyTorch in distributed training contexts
Strong understanding of networking, storage, and compute resource management for ML workloads
Experience developing APIs and managing distributed systems for both batch and real-time workloads
Solid debugging and monitoring skills with expertise in observability tools for containerized environments
Preferred Skills Experience with Kubernetes operators and custom controllers for ML workloads
Advanced Slurm administration including multi-cluster federation and advanced scheduling policies
Familiarity with GPU cluster management and CUDA optimization
Experience with other ML frameworks like TensorFlow or distributed training libraries
Background in HPC environments, parallel computing, and high-performance networking
Knowledge of infrastructure as code (Terraform, Ansible) and GitOps practices
Experience with container registries, image optimization, and multi-stage builds for ML workloads
Required Experience Demonstrated experience managing large-scale Kubernetes deployments in production environments
Proven track record with Slurm cluster administration and HPC workload management
Previous roles in SRE, DevOps, or Platform Engineering with focus on ML infrastructure
Experience supporting both long-running training jobs and high-availability inference services
Ideally, 3-5 years of relevant experience in ML systems deployment with specific focus on cluster orchestration and resource management
The cash compensation range for this role is $190,000 - $250,000.
Final offer amounts are determined by multiple factors, including, experience and expertise, and may vary from the amounts listed above.
Equity: In addition to the base salary, equity may be part of the total compensation package.
Benefits: Comprehensive health, dental, and vision insurance for you and your dependents. Includes a 401(k) plan.
Create a Job Alert
Interested in building your career at Perplexity AI? Get future opportunities sent straight to your email.
Apply for this job indicates a required field
First Name *
Last Name *
Email *
Phone *
Resume/CV *
Enter manually
Accepted file types: pdf, doc, docx, txt, rtf
Enter manually
Accepted file types: pdf, doc, docx, txt, rtf
Website
LinkedIn Profile
Will you now or in the future require visa sponsorship for employment? * Select...
Perplexity has an office-centric work model with 4 days per week in the office from the San Francisco Bay Area or New York City. Are you willing to come in 4 days per week? * Select...
If you are not based in any of these locations, are you open to relocation to San Francisco, Palo Alto, or New York City? * Select...
What are you looking for in your next role? *
#J-18808-Ljbffr
- Location:
- New York, NY, United States
- Salary:
- $200,000 - $250,000
- Category:
- IT & Technology