Vacancy is archived. Applications are no longer accepted.

Senior Solutions Architect, Gen AI

at Nvidia
πŸ“ World
πŸ“ United States
SENIOR
βœ… Remote βœ… On-site

SCRAPED

Used Tools & Technologies

Not specified

Required Skills & Competences ?

Docker @ 4 Kubernetes @ 4 Linux @ 4 DevOps @ 4 GCP @ 4 TensorFlow @ 4 AWS @ 4 Azure @ 4 Communication @ 4 LLM @ 4 PyTorch @ 4 GPU @ 4

Details

NVIDIA is seeking a Solutions Architect skilled in full-cycle Generative AI development and deployment to support customers adopting NVIDIA's full-stack technologies. As part of the Hospitality and Travel Solutions Architecture team you will work with hotel and transportation customers to integrate accelerated computing and AI into their stacks. You will engage directly with developers, researchers, and data scientists at strategic customers and partner with internal engineering, product, and sales teams to drive product strategy and design wins.

Responsibilities

  • Provide hands-on technical mentorship to partners and customers on the NVIDIA GenAI stack.
  • Guide customers to develop and deploy agentic AI workflows on NVIDIA platforms, quantifying benefits of accelerated computing software and hardware.
  • Build demonstrations and POCs that address critical customer business needs and draft requirements for missing features to unblock progress.
  • Educate customers on new NVIDIA GenAI technologies and platforms; prepare and deliver technical training sessions and workshops.
  • Create collateral (notebooks, blogs) applied to industry use-cases.
  • Partner with NVIDIA engineering, product, and sales teams to secure design wins and enable development of NVIDIA features through customer feedback and POC evaluations.

Requirements

  • Master's or Ph.D. in Computer Science, Artificial Intelligence, or equivalent experience.
  • 8+ years of hands-on experience in a technical AI role, with a strong emphasis on Generative AI.
  • Proficiency with current model architectures and the ability to articulate computational complexities of each architecture.
  • Proven track record of deploying and optimizing large language models (LLMs) for inference in production using inferencing engines such as vLLM, TRT-LLM, SGLang, etc.
  • Expertise in training and fine-tuning LLMs using frameworks such as TensorFlow, PyTorch, or Hugging Face Transformers.
  • Solid understanding of GPU cluster architecture and experience leveraging parallel processing for accelerated model training and inference.
  • Experience with DevOps and related tooling (Docker, Kubernetes, GitLab, Linux command line, Shell).
  • Excellent communication and teamwork skills; ability to explain complex technical concepts to technical and non-technical stakeholders.
  • Experience leading workshops, delivering trainings, and presenting technical solutions.

Ways to Stand Out

  • Experience with agentic AI frameworks, tools, and protocols (e.g., LangChain, LangGraph, MCP).
  • Understanding of multimodal LLMs and VLMs.
  • Experience deploying LLMs at scale on mainstream cloud providers (AWS, Azure, GCP).
  • Proven ability to profile and optimize inference latency, throughput, memory, and I/O utilization.
  • Strong mathematical understanding of parallelization techniques in Gen AI.

Compensation & Benefits

  • Base salary ranges by level (location, experience, and comparable pay considered):
    • Level 4: 184,000 USD - 287,500 USD
    • Level 5: 224,000 USD - 356,500 USD
  • Eligible for equity and additional benefits (see NVIDIA benefits).

Additional Information

  • Applications accepted at least until July 29, 2025.
  • NVIDIA is an equal opportunity employer committed to diversity and inclusion.