Senior Product Manager - Inference Benchmarking

at Nvidia
USD 208,000-327,800 per year
SENIOR
βœ… Hybrid

SCRAPED

Used Tools & Technologies

Not specified

Required Skills & Competences ?

Marketing @ 4 Distributed Systems @ 4 Planning @ 4 Agile @ 7

Details

NVIDIA is at the center of Deep Learning, Artificial Intelligence, and Autonomous Vehicles. This role focuses on building benchmarks, products, and tools that help customers understand the true performance and total cost of ownership (TCO) of AI inference workloads. The product manager will operate across NVIDIA to define what to measure, how to measure it, and how to present results that help customers balance performance, cost, and user experience for inference deployments.

Responsibilities

  • Serve as a Subject Matter Expert on AI Inference: maintain a deep understanding of the entire inference stack, including performance, scaling across workloads, and emerging technologies (for example, disaggregated serving), to guide technical and product strategy.
  • Drive product strategy through market and user insights: conduct research on market trends, competitor activities, and customer feedback; translate insights into actionable development plans and product ideas.
  • Lead partner collaboration and project execution: manage external partnerships, lead project planning, define tasks and deliverables, act as primary liaison, and educate partners on product value to ensure alignment and timely issue resolution.
  • Spearhead cross-functional product introduction: collaborate with engineering, design, operations, sales, and marketing to drive new product introduction and transitions from conception to launch.
  • Enhance product and user experience: contribute to ideation, design, and testing of user experiences, ensuring products meet customer and partner expectations.

Requirements

  • BS or MS in Computer Science, Computer Engineering, or a related field (or equivalent experience).
  • 12+ years of product-management experience in enterprise technology.
  • Subject matter expertise in AI inference and the inference stack; understanding of how components perform and scale across workloads.
  • Ability to articulate trade-offs among latency, throughput, cost, and reliability to both engineering and executive audiences.
  • Strong cross-functional execution skills: writes clear specifications and PRDs, produces GTM collateral, and leads agile processes.

Ways to stand out:

  • Masters/PhD or expertise in distributed systems.
  • Demonstrated experience in inference (within the stack or ecosystem) and integrating with enterprise platforms; deployments at modern data-center scale.

Benefits

  • Base salary range: 208,000 USD - 327,750 USD (final base salary determined by location, experience, and comparable pay).
  • Eligible for equity and company benefits.
  • NVIDIA emphasizes growth, innovation, and a diverse work environment; equal opportunity employer.

Additional info:

  • Location: Santa Clara, CA, United States.
  • Office policy: Hybrid (#LI-Hybrid).
  • Applications accepted at least until September 9, 2025.