Principal Artificial Intelligence Algorithms Engineer
at Nvidia
π Santa Clara, United States
USD 272,000-425,500 per year
SCRAPED
Used Tools & Technologies
Not specified
Required Skills & Competences ?
Python @ 6 Algorithms @ 4 Machine Learning @ 4 Hiring @ 4 Debugging @ 6 LLM @ 4 PyTorch @ 4 GPU @ 4Details
NVIDIA is hiring for the core AI Frameworks team (Megatron Core and NeMo Framework) to design, develop, and optimize diverse real-world workloads for Large Language Models (LLM) and Multimodal (MM) foundation models. The role focuses on expanding framework capabilities, designing and implementing distributed training algorithms and model-parallel paradigms, optimizing performance on NVIDIA GPU architectures, and collaborating with internal partners and the open source community.
Responsibilities
- Develop algorithms for AI/Deep Learning, data analytics, machine learning, or scientific computing.
- Contribute to and advance open source Megatron Core and NeMo Framework.
- Solve large-scale, end-to-end AI training and inference challenges across the full model lifecycle: orchestration, data pre-processing, training, tuning, and deployment.
- Work at the intersection of computer architecture, libraries, frameworks, AI applications and the full software stack.
- Innovate and improve model architectures, distributed training algorithms, and model-parallel paradigms.
- Perform performance tuning and optimizations, model training and fine-tuning with mixed precision recipes on next-generation NVIDIA GPU architectures.
- Research, prototype, and develop robust and scalable AI tools and pipelines.
Requirements
- MS, PhD or equivalent experience in Computer Science, AI, Applied Math, or related fields and 10+ years of industry experience.
- Experience with AI frameworks (e.g., PyTorch, JAX) and/or inference and deployment environments (e.g., TRTLLM, vLLM, SGLang).
- Proficient in Python programming, software design, debugging, performance analysis, test design and documentation.
- Consistent record of working effectively across multiple engineering initiatives and improving AI libraries with new innovations.
- Strong understanding of AI/Deep-Learning fundamentals and their practical applications.
Ways to Stand Out
- Hands-on experience in large-scale AI training with deep understanding of core compute system concepts (latency/throughput bottlenecks, pipelining, multiprocessing) and demonstrated excellence in performance analysis and tuning.
- Expertise in distributed computing, model parallelism, and mixed precision training.
- Prior experience with Generative AI techniques applied to LLM and Multi-Modal learning (text, image, video).
- Knowledge of GPU/CPU architecture and related numerical software.
- Contributions to open source deep learning frameworks.
Benefits and Other Information
- Base salary range: 272,000 USD - 425,500 USD (determined based on location, experience, and pay of employees in similar positions).
- Eligible for equity and benefits.
- Applications accepted at least until August 15, 2025.
- NVIDIA is an equal opportunity employer and values diversity in hiring and promotion practices.