Solutions Architect, AI Infrastructure

at Nvidia
USD 148,000-235,800 per year
MIDDLE
βœ… Remote βœ… Hybrid

SCRAPED

Used Tools & Technologies

Not specified

Required Skills & Competences ?

Docker @ 3 Kubernetes @ 3 Linux @ 3 DevOps @ 3 MLOps @ 3 Networking @ 3 Product Management @ 3 Debugging @ 3 CUDA @ 3 GPU @ 3

Details

NVIDIA is looking for an experienced systems and network infrastructure Solutions Architect to join the Solution Architecture Engineering (SA) team focused on supporting accelerated computing applications. You will drive end-to-end technology solutions integration with strategic customers and offer recommendations to business and engineering teams on product technology. The role involves working on data center GPU server and networking infrastructure deployments, building POCs and demos, and supporting bring-up and debug of clusters and networks.

Responsibilities

  • Work with NVIDIA Consumer Internet and IT Services customers on data center GPU server and networking infrastructure deployments as a solution architect. Guide customer discussions on network topologies, compute/storage and support bring up of server/network/cluster deployments. On-site visits to customer data centers are required during bring up phases.
  • Identify new project opportunities for NVIDIA products and technology solutions in data center and AI applications. Work closely with Systems/Network Engineering, Product Management and Sales teams.
  • Act as a trusted technical advisor to customers by conducting regular technical meetings for product roadmaps, cluster debugging, feature discussions and introduction to new technology solutions.
  • Build custom product demonstrations and POCs addressing critical customer business needs.
  • Analyze and debug compute and network performance issues.

Requirements

  • BS/MS/PhD in Electrical/Computer Engineering, Computer Science, Physics, or other engineering fields, or equivalent experience.
  • Ideally 5+ years of Solution Engineering (or similar engineering roles) experience, with motivation and skills to drive data center engagements.
  • System-level understanding of server architecture, NICs, Linux, system software and kernel drivers.
  • Practical knowledge of networking β€” switching & routing for Ethernet/InfiniBand β€” and data center infrastructure (power/cooling).
  • Knowledge of DevOps/MLOps technologies such as Docker/containers and Kubernetes is a plus.
  • Effective time management and the ability to balance multiple tasks.
  • Ability to clearly communicate ideas and code via documents and presentations.

Ways To Stand Out From The Crowd

  • External customer-facing experience and background.
  • Experience with bring up and deployment of large clusters.
  • Systems engineering, coding, and debugging skills including experience with C/C++, Linux kernel and drivers.
  • Hands-on experience with NVIDIA systems/SDKs (e.g., CUDA), NVIDIA networking technologies (e.g., DPU, RoCE, InfiniBand), and/or ARM CPU solutions.
  • Familiarity with virtualization technology concepts.

Benefits

  • Eligible for equity and company benefits. (Link to NVIDIA benefits provided in original posting.)

Compensation

  • Base salary range: 148,000 USD - 235,750 USD. Final base salary will be determined based on location, experience, and pay of employees in similar positions.

Location & Travel

  • Location listed: US, CA, Santa Clara. The team makes extensive use of conferencing tools, is open to remote work locations, and requires occasional (~20%) travel for on-site customer visits and industry events.

Equal Opportunity

  • NVIDIA is committed to fostering a diverse work environment and is an equal opportunity employer. Applications accepted at least until August 14, 2025.