We are looking for Research Scientists passionate about systems for large generative models! We are searching for world-class researchers in systems for deep learning to join our applied deep learning research team which pioneered Megatron and MT-NLG. Our team designs and builds the systems used to train and deploy state-of-the-art large language and multimodal foundation models by employing techniques from both supercomputing and distributed computing. We collaborate with both internal and external partners, including hardware and software teams, to train and serve some of the world’s largest and best foundation models as efficiently as possible. If you are passionate about the latest research and technologies revolutionizing generative AI and want to explore creative new paradigms for training and serving models on the latest and best compute platforms and clusters available, this team will be a great fit for you. What you'll be doing: Tackle large-scale distributed systems capable of performing end-to-end AI training and inference. Develop, design, and implement algorithms to accelerate training and inference of large language models (LLMs) and large multi-modal models. Work with deep learning hardware teams to co-design hardware and software for future model architectures. Work closely with product and hardware architecture teams to integrate research and developments into products. Contribute to open-source software and publish at top conferences.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Education Level
Ph.D. or professional degree