Help Luma build some of the biggest & fastest AI supercomputing clusters in the world! As a High-Performance Computing engineer, you’ll work at the intersection of hardware and software, designing systems that deliver the maximum possible performance for running large-scale AI models. We work at the very cutting edge of speed and scale, combining the traditions of High-Performance Computing (HPC) in a modern cloud environment.
For this role, it’s important you understand how to combine CPU’s, GPU’s, and network devices into systems that are then deployed at a large scale to peak efficiency. You understand the lowest levels of the software platforms that sit on top of this hardware, including how to best optimize the Linux kernel and user-space code. You are capable of writing code to automate the monitoring and healing of these systems, commanding a large number of servers with few people.
Responsibilities
In this role, you will work closely with and directly accelerate machine learning researchers, but don't need to be a machine learning expert yourself.
We value people who can quickly obtain a deep technical understanding of new domains and enjoy being self-directed and identifying the most important problems to solve.
You’ll be managing training HPC clusters at Luma from provisioning to performance tuning.
Areas of work will include observability, distributed job tracing, GPU diagnostics, software environment management and additional tooling plus work on the actual code to enable necessary features.
We believe that increasing compute is a huge lever to AI progress. You will have a direct impact on our ability to grow to an unprecedented scale and likewise produce unprecedented results.
Experience
8+ years experience as infrastructure engineer or Devops in large and complex distributed systems.
Deep understanding of networking, bonus points for experience in HPC networking.
Experience developing high-quality software in a general-purpose programming language, preferably including Python.
Excellent problem-solving skills and attention to detail.
Experience with GPUs in large scale clusters is strongly preferred.
Strong knowledge of observability and monitoring in distributed systems.
Tenacious at troubleshooting hardware and network topology failures in distributed systemsIndependently driven and able to own problems and build solutions from end-to-end.
Experience with large scale data center operations, proficiency in cloud orchestration and system tools.
Compensation
In addition to cash base pay, you'll also receive a sizable grant of Luma's equity.
The pay range for this position is $180000- 220000/yr for Bay Area. Base pay offered will vary depending on job-related knowledge, skills, candidate location, and experience.
Your application is reviewed by real people.
At Luma AI, we believe that multimodality is critical for intelligence. To go beyond language models and build more aware, capable and useful systems, the next step function change will come from vision. So, we are working on training and scaling up multimodal foundation models for systems that can see and understand, show and explain, and eventually interact with our world to effect change.
We will deploy these systems to make a new kind of intelligent creative partner that can imagine with us. Free and away from the pressure of being creative. It's for all of us whose imaginations have been constrained, who've had to channel vivid dreams through broken words, hoping others will see what we see in our mind's eye. A partner that can help us show — not just tell.
Dream Machine is an early step to building that. Try it here
Luma is bringing together the best team in the world to achieve our goal, from researchers to engineers and designers to growth operators
Luma is not just a lab - we are deeply product focused and our vision merging AI models and delightful products is unique in the industry
We build. We ship. Our early products have been wildly successful