About The Position

NVIDIA is seeking capable customer-facing hardware engineers to work directly with Cloud Scale Providers (CSP’s) deploying next generation AI-centric Data Centers. The HW Systems Engineer is front-and-center in deploying next generation Nvidia MGX platforms, such as Grace-Blackwell and Vera Ruben NVL72 racks, at our largest customers. As AI continues to evolve into the era of reasoning, Data Centers focused solely on AI, known as AI Factories, are vital to scale compute and networking infrastructure needed for agentic AI processing. The CSP HW Systems Engineer is the central point of contact for the team of Applications Engineers deploying these AI Factories. The ideal candidate is proactive, customer friendly, and able to lead large cross-functional teams to debug highly complex problems. The CSP Systems Engineering team is deploying the latest generation racks, so close collaboration with architecture and NPI engineering teams is required.

Requirements

  • Bachelor's or Master's degree in Computer Engineering, Electrical Engineering, or a related field (or equivalent experience).
  • 5+ years of proven experience in system-level design and integration of server products from concept to deployment.
  • Solid understanding of x86 server architecture, PCIe, DDR, Infiniband and high-speed interconnects
  • Basic understanding of BMC (Baseboard Management Controller) architecture, I2C (SMBus), power management and system telemetry controls
  • Familiarity with Linux OS and command line experience
  • Knowledge of the latest PCIe Gen5 and Gen6 technical interface challenges
  • Strong problem-solving and analytical skills as well as excellent communication and teamwork skills
  • Strong analytical, problem-solving, time-management, and organizational skills, with the ability to manage multiple complex initiatives in dynamic environments.

Nice To Haves

  • Self-motivated and eager to learn.
  • Can work under high pressure and dynamic environments.
  • In-depth understanding of CPU, GPU, Networking and architectural tradeoffs is a plus!
  • A desire to understand technology deeply.
  • Can communicate and explain clearly information relevant to your audience.

Responsibilities

  • Collaborate with major cloud service providers (CSP), their OEM/ODM’s, and internal teams to help deploy the latest Nvidia Vera-Ruben AI rack’s
  • Work with other domain expert teams at NVIDIA to ensure customer solutions are optimized for the highest performance servers in the world
  • Solve deep server system technical issues at the hardware, software and application level, ensuring customer success and time to market
  • Act as the central point of contact between CSP customers and Nvidia architecture teams to develop future GPU-accelerated data center architectures and roadmaps
  • Excellent verbal, written communication, and technical presentation skills in English.
  • Up to 30% travel expected

Benefits

  • NVIDIA offers highly competitive salaries and a comprehensive benefits package.
  • You will also be eligible for equity and benefits .
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service