Forward Deployed Engineer, AI Accelerator
at Nvidia
π Santa Clara, United States
USD 168,000-322,000 per year
SCRAPED
Used Tools & Technologies
Not specified
Required Skills & Competences ?
Ansible @ 3 Docker @ 3 Kubernetes @ 3 Linux @ 6 DevOps @ 6 Terraform @ 3 Python @ 3 GCP @ 3 CI/CD @ 3 Distributed Systems @ 3 MLOps @ 3 TensorFlow @ 3 AWS @ 3 Azure @ 3 ServiceNow @ 3 API @ 3 PyTorch @ 3 Salesforce @ 3 CUDA @ 3 GPU @ 3Details
NVIDIA is seeking a Forward Deployed Engineer to join the AI Accelerator team, working directly with strategic customers to implement and optimize pioneering AI workloads. You will provide hands-on technical support for advanced AI implementations and complex distributed systems, helping customers achieve optimal performance from NVIDIA's AI platform across diverse environments.
Responsibilities
- Design and deploy custom AI solutions including distributed training, inference optimization, and MLOps pipelines across customer environments.
- Provide remote technical support to strategic customers: optimize AI workloads, diagnose and resolve performance issues, and guide technical implementations through virtual collaboration.
- Deploy and manage AI workloads across DGX Cloud, customer data centers, and CSP environments using Kubernetes, Docker, and GPU scheduling systems.
- Profile and optimize large-scale model training and inference workloads, implement monitoring solutions, and resolve scaling challenges.
- Build custom integrations with customer systems, develop APIs and data pipelines, and implement enterprise software connections.
- Create implementation guides, documentation for resolution approaches, and standard methodologies for complex AI deployments.
Requirements
- 8+ years of experience in customer-facing technical roles (Solutions Engineering, DevOps, ML Infrastructure Engineering) or equivalent.
- BS, MS, or Ph.D. in Computer Science, Computer Engineering, Electrical Engineering (or related technical field) or equivalent experience.
- Strong proficiency with Linux systems, distributed computing, Kubernetes, and GPU scheduling.
- AI/ML experience supporting inference workloads and training at large scale.
- Programming skills in Python, with experience in PyTorch, TensorFlow, or similar AI frameworks.
- Ability to engage with customers and work effectively with technical teams under high-pressure situations.
Ways to stand out
- NVIDIA ecosystem experience with DGX systems, CUDA, NeMo, Triton, or NIM.
- Hands-on experience with cloud AI platforms (AWS, Azure, GCP).
- MLOps expertise: containerization, CI/CD pipelines, and observability tooling.
- Infrastructure-as-code experience with Terraform, Ansible, or similar automation tools.
- Enterprise software integration experience with platforms such as Salesforce or ServiceNow.
Compensation & Benefits
- Base salary ranges (location, experience, and level dependent):
- Level 4: 168,000 USD - 264,500 USD per year
- Level 5: 200,000 USD - 322,000 USD per year
- Eligible for equity and a generous benefits package. See NVIDIA benefits for details.
Additional information
- Location: Santa Clara, CA, United States. #LI-Hybrid
- Employment type: Full time
- Applications for this job will be accepted at least until October 13, 2025.
- NVIDIA is an equal opportunity employer committed to fostering a diverse work environment.