Our teams' mission is to explore, develop and help productionize high performance software & hardware technologies for AI at datacenter scale. We achieve this via concurrent design and optimization of many aspects of the system from models and runtime all the way to the AI hardware, optimizing across compute, network and storage.
What you'll do
Explore, co-design and optimize parallelisms, compute efficiency, distributed training/inference paradigms and algorithms to improve the scalability, efficiency and reliability of inference and large-scale training systems.
What you need
- Currently has, or is in the process of obtaining a PhD degree in Computer Science, Computer Vision, Generative AI, NLP, relevant technical field, or equivalent practical experience.