Senior Inference Technical Product Marketing Manager - Accelerated Computing

at Nvidia
USD 144,000-287,500 per year
SENIOR
βœ… Hybrid

SCRAPED

Used Tools & Technologies

Not specified

Required Skills & Competences ?

Marketing @ 4 Software Development @ 4 TensorFlow @ 3 Leadership @ 6 Communication @ 4 Networking @ 4 Product Management @ 4 LLM @ 7 PyTorch @ 3 CUDA @ 4

Details

We are looking for a Senior Technical Product Marketing Manager to work in our rapidly growing data center business, focused on inference marketing. The role partners closely with engineering to understand and communicate the technical capabilities of the inference stack across GPUs, CPUs, networking, CUDA libraries, model architectures and deployment techniques (parallelisms, configurations, etc.). You will influence NVIDIA's technical marketing strategy to showcase leadership in AI inference.

Responsibilities

  • Drive inference platform technical go-to-market efforts.
  • Work closely with engineering and product management to understand technical capabilities of the inference stack (GPUs, CPUs, networking, CUDA libraries, model architectures, deployment techniques such as parallelisms and configurations).
  • Monitor and review model architectures, frameworks, arXiv papers, whitepapers, and deployment techniques (e.g., disaggregated serving, KV cache implementations) to identify intersections between the latest AI models and NVIDIA's platform to maximize performance and minimize TCO.
  • Develop clear positioning, messaging and assets (blogs, whitepapers, presentations, analyst briefings, conference seminars) to highlight NVIDIA's leadership in inference.
  • Track competitive inference announcements and prepare technical/business responses for internal and developer audiences.
  • Assist in building keynote slides for executives in areas where you are a subject matter expert.

Requirements

  • BS in Computer Science, Engineering, or related field (or equivalent experience); Masters preferred.
  • 6+ years experience in LLM/AI/ML development in an engineering role followed by 5+ years in product management or technical product marketing of AI/ML products.
  • Deep understanding of modern data center architectures, accelerated computing, distributed inference, and deep learning frameworks.
  • Familiarity with frameworks and inference-specific optimizations such as PyTorch, TensorFlow, JAX, Dynamo, Triton Inference Server, TensorRT-LLM, vLLM, SGLang.
  • Experience with deployment topics such as disaggregated serving and KV cache implementations.
  • Market awareness and experience conducting technical competitive analysis.
  • Strong cross-functional collaboration and influence across engineering, product management, sales, and marketing.
  • Excellent communication, asset creation, and storytelling skills to translate technical concepts for technical and business audiences.
  • Ability to present to executive audiences.

Ways to Stand Out

  • Hands-on experience with AI inferencing workflows using NVIDIA or open-source serving frameworks on accelerated computing in the data center.
  • Experience developing LLM models.
  • Experience working with hyperscale cloud providers.
  • Background in software development, AI infrastructure, or data center silicon.
  • Demonstrated ability to engage with executive leadership and external partners.
  • Published technical content or speaking experience at industry events and a portfolio of published marketing/launch assets.

Compensation & Benefits

  • Base salary ranges provided by level:
    • Level 4: 144,000 USD - 230,000 USD
    • Level 5: 184,000 USD - 287,500 USD
  • Eligible for equity and benefits.

Location & Employment Type

  • Location: Santa Clara, CA, United States.
  • Employment type: Full time.
  • Office policy: Hybrid (#LI-Hybrid).

Additional Information

  • Applications accepted at least until July 29, 2025.
  • NVIDIA is committed to diversity and is an equal opportunity employer.