Senior Deep Learning Engineer

at Nvidia
USD 148,000-287,500 per year
SENIOR
✅ On-site

SCRAPED

Used Tools & Technologies

Not specified

Required Skills & Competences ?

Software Development @ 6 Algorithms @ 4 Mathematics @ 4 PyTorch @ 6

Details

We are at the forefront of advancing the capabilities of artificial intelligence and are seeking a Senior Deep Learning Engineer to contribute to next-generation inference optimizations and deliver industry-leading performance without compromising model quality. The role focuses on scaling test-time compute and optimizing low-latency inference through cross-stack optimizations at algorithmic and system levels, collaborating with research, framework development, and silicon architecture teams.

Responsibilities

  • Keep abreast of the latest advancements in generative AI research.
  • Prototype and analyze emergent techniques in the test-time compute space such as output refinement, speculation, and retrieval; identify opportunities for algorithmic and system optimizations.
  • Pioneer development of innovative optimizations to enable high-quality inferencing on NVIDIA GPUs.
  • Collaborate closely with production teams to incorporate the latest advancements into cutting-edge software frameworks.
  • Work across algorithms, software and performance teams to deliver production-ready, high-quality solutions.

Requirements

  • Master's degree (or equivalent experience) in Computer Science, Artificial Intelligence, Applied Mathematics, or a related field.
  • Strong foundation in deep learning, with particular emphasis on generative models and inferencing.
  • At least 5 years of relevant software development experience in modern deep learning frameworks such as PyTorch.
  • Growth mindset and pragmatic attitude.
  • Familiarity with how computer architecture relates to AI algorithm development and with cross-stack (algorithm + system) optimization approaches.

Ways to Stand Out

  • Published research or noteworthy contributions to deep learning, particularly in inference-time compute, conditional compute, speculative decoding, etc.
  • Experience prototyping and/or deploying emergent test-time compute techniques.
  • Experience collaborating across algorithms, software and performance teams.
  • Familiarity with computer architecture and its relation to AI algorithm development.

Compensation & Benefits

  • Base salary range: 148,000 USD - 235,750 USD for Level 3, and 184,000 USD - 287,500 USD for Level 4. Actual base salary will be determined based on location, experience, and pay of employees in similar positions.
  • Eligibility for equity and other benefits (see company benefits link).

Additional Notes

  • The role is embedded in an interdisciplinary environment supporting AI datacenter efficiency, hardware-aware software, and system software development.
  • Applications accepted at least until July 29, 2025.