Forward Deployed Engineer, AI Accelerator

at Nvidia
USD 168,000-322,000 per year
MIDDLE
βœ… Hybrid

SCRAPED

Used Tools & Technologies

Not specified

Required Skills & Competences ?

Ansible @ 3 Docker @ 3 Kubernetes @ 3 Linux @ 6 DevOps @ 6 Terraform @ 3 Python @ 3 GCP @ 3 CI/CD @ 3 Distributed Systems @ 3 MLOps @ 3 TensorFlow @ 3 AWS @ 3 Azure @ 3 ServiceNow @ 3 API @ 3 PyTorch @ 3 Salesforce @ 3 CUDA @ 3 GPU @ 3

Details

NVIDIA is seeking a Forward Deployed Engineer to join the AI Accelerator team, working directly with strategic customers to implement and optimize pioneering AI workloads. You will provide hands-on technical support for advanced AI implementations and complex distributed systems, helping customers achieve optimal performance from NVIDIA's AI platform across diverse environments.

Responsibilities

  • Design and deploy custom AI solutions including distributed training, inference optimization, and MLOps pipelines across customer environments.
  • Provide remote technical support to strategic customers: optimize AI workloads, diagnose and resolve performance issues, and guide technical implementations through virtual collaboration.
  • Deploy and manage AI workloads across DGX Cloud, customer data centers, and CSP environments using Kubernetes, Docker, and GPU scheduling systems.
  • Profile and optimize large-scale model training and inference workloads, implement monitoring solutions, and resolve scaling challenges.
  • Build custom integrations with customer systems, develop APIs and data pipelines, and implement enterprise software connections.
  • Create implementation guides, documentation for resolution approaches, and standard methodologies for complex AI deployments.

Requirements

  • 8+ years of experience in customer-facing technical roles (Solutions Engineering, DevOps, ML Infrastructure Engineering) or equivalent.
  • BS, MS, or Ph.D. in Computer Science, Computer Engineering, Electrical Engineering (or related technical field) or equivalent experience.
  • Strong proficiency with Linux systems, distributed computing, Kubernetes, and GPU scheduling.
  • AI/ML experience supporting inference workloads and training at large scale.
  • Programming skills in Python, with experience in PyTorch, TensorFlow, or similar AI frameworks.
  • Ability to engage with customers and work effectively with technical teams under high-pressure situations.

Ways to stand out

  • NVIDIA ecosystem experience with DGX systems, CUDA, NeMo, Triton, or NIM.
  • Hands-on experience with cloud AI platforms (AWS, Azure, GCP).
  • MLOps expertise: containerization, CI/CD pipelines, and observability tooling.
  • Infrastructure-as-code experience with Terraform, Ansible, or similar automation tools.
  • Enterprise software integration experience with platforms such as Salesforce or ServiceNow.

Compensation & Benefits

  • Base salary ranges (location, experience, and level dependent):
    • Level 4: 168,000 USD - 264,500 USD per year
    • Level 5: 200,000 USD - 322,000 USD per year
  • Eligible for equity and a generous benefits package. See NVIDIA benefits for details.

Additional information

  • Location: Santa Clara, CA, United States. #LI-Hybrid
  • Employment type: Full time
  • Applications for this job will be accepted at least until October 13, 2025.
  • NVIDIA is an equal opportunity employer committed to fostering a diverse work environment.