Used Tools & Technologies
Not specified
Required Skills & Competences ?
Marketing @ 3 Kubernetes @ 3 Python @ 6 Algorithms @ 6 Data Structures @ 6 JavaScript @ 6 Microservices @ 3 Debugging @ 6 API @ 3 QA @ 3 Engineering Management @ 3 LLM @ 3Details
NVIDIA has been transforming computer graphics, PC gaming, and accelerated computing for more than 25 years. Today NVIDIA is tapping into the unlimited potential of AI to define the next era of computing. As a System Software Engineer - AI and Cloud you will work on cloud-native, full-stack AI systems, contributing to developer-facing content, technical reports, and product improvements. This role is located in Silicon Valley (Santa Clara, CA). Applications accepted at least until August 26, 2025. You will be eligible for equity and benefits.
Responsibilities
- Evaluate cloud-native, full-stack applications using microservices architecture to power AI use cases, leveraging NVIDIA frameworks, SDKs, and microservices.
- Design and implement agentic workflows using advanced techniques such as Retrieval-Augmented Generation (RAG) and modern AI models.
- Evaluate user experiences and analyze technical performance of AI solutions; compile findings into comprehensive reports and provide product improvement suggestions to senior executives and engineering management.
- Collaborate with product, marketing, hardware, software engineering, and QA teams to improve NVIDIA's product offerings.
- Develop developer-focused content (tutorials, code samples) demonstrating NVIDIA tools and libraries.
- Write technical whitepapers and product briefs, and run technical demos at industry conferences.
Requirements
- Bachelor's or Master's degree in Software Engineering, Computer Science, Computer Engineering, Electrical Engineering, or a related field, or equivalent experience.
- 3+ years of industry experience.
- Proficiency in Python and JavaScript for programming and debugging; strong foundation in data structures, algorithms, and software design principles.
- Basic familiarity with C++ and its application in high-performance computing environments.
- Experience crafting cloud-native systems optimized for Kubernetes deployment, using inference frameworks such as vLLM and NVIDIA Triton Inference Server.
- Solid understanding of API design principles for building scalable, production-ready inference systems.
Ways To Stand Out / Preferred Qualifications
- Advanced knowledge of LLMs, modern AI software architecture, and cloud APIs.
- Contributions to public-facing technical content and open-source projects.
- Expertise deploying LLM inference frameworks like Triton Inference Server, vLLM, or TensorRT, including on Kubernetes or edge devices to improve performance.
Compensation & Benefits
- Base salary ranges by level:
- Level 2: 120,000 USD - 189,750 USD
- Level 3: 148,000 USD - 235,750 USD
- Eligible for equity and NVIDIA benefits.
Other Information
- Location: Santa Clara, CA, United States (Silicon Valley).
- Role type: Full time.
- NVIDIA is an equal opportunity employer committed to fostering a diverse work environment.