Senior Research Engineer
at Nvidia
π Santa Clara, United States
USD 160,000-299,000 per year
SCRAPED
Used Tools & Technologies
Not specified
Required Skills & Competences ?
Python @ 6 Algorithms @ 4 Machine Learning @ 6 Debugging @ 6 LLM @ 4 PyTorch @ 4 GPU @ 4Details
Join NVIDIA and help build the software that will define the future of generative AI. You will work at the intersection of research and engineering to create the next-generation post-training software stack for Nemotron models, collaborating with the Post-Training and Frameworks teams and contributing to open-source projects such as NeMo-RL, Megatron Core, and the NeMo Framework.
Responsibilities
- Work with applied researchers to design, implement, and test next-generation RL and post-training algorithms.
- Contribute to and advance open-source projects (NeMo-RL, Megatron Core, NeMo Framework, and other NVIDIA software).
- Be engaged as part of a team during Nemotron models post-training.
- Solve large-scale, end-to-end AI training and inference challenges across the full model lifecycle: orchestration, data pre-processing, model training and tuning, and model deployment.
- Work across computer architecture, libraries, frameworks, AI applications, and the full software stack.
- Perform performance tuning and optimizations; implement model training with mixed precision recipes on next-generation NVIDIA GPU architectures.
- Publish and present results at academic and industry conferences.
Requirements
- BS, MS, or PhD in Computer Science, AI, Applied Math, or related fields, or equivalent experience.
- 3+ years of proven experience in machine learning, systems, distributed computing, or large-scale model training.
- Experience with AI frameworks such as PyTorch or JAX.
- Experience with at least one inference/deployment environment such as vLLM, SGLang, or TRT-LLM.
- Proficient in Python programming, software design, debugging, performance analysis, test design, and documentation.
- Strong understanding of AI/deep-learning fundamentals and practical applications.
Ways to Stand Out / Additional Qualifications
- Contributions to open-source deep learning libraries.
- Hands-on experience in large-scale AI training and deep understanding of compute system concepts (latency/throughput bottlenecks, pipelining, multiprocessing) with demonstrated excellence in performance analysis and tuning.
- Expertise in distributed computing, model parallelism, and mixed precision training.
- Prior experience with generative AI techniques applied to LLMs and multi-modal learning (text, image, video).
- Knowledge of GPU/CPU architecture and related numerical software.
Compensation & Benefits
- Base salary range: Level 3 β 160,000 USD to 258,750 USD; Level 4 β 184,000 USD to 299,000 USD. Actual base salary will be determined based on location, experience, and pay of employees in similar positions.
- Eligible for equity and benefits.
Other
- Applications accepted until October 13, 2025.
- NVIDIA is an equal opportunity employer and is committed to fostering a diverse work environment.