Software Engineer, Perception Optimization

WaymoSan Francisco, CA
2dHybrid

About The Position

Waymo is an autonomous driving technology company with the mission to be the world's most trusted driver. Since its start as the Google Self-Driving Car Project in 2009, Waymo has focused on building the Waymo Driver—The World's Most Experienced Driver™—to improve access to mobility while saving thousands of lives now lost to traffic crashes. The Waymo Driver powers Waymo’s fully autonomous ride-hail service and can also be applied to a range of vehicle platforms and product use cases. The Waymo Driver has provided over ten million rider-only trips, enabled by its experience autonomously driving over 100 million miles on public roads and tens of billions in simulation across 15+ U.S. states. The Hardware Abstraction Layer (HAL) Optimization team uses our know-how of CPUs, GPUs, and clever algorithms to process Waymo's sensor data super-fast and feed it to the Perception models. We specialize in low-level understanding of compute hardware, and we apply that knowledge wherever we can, throughout the whole stack. We are looking for engineers with a CPU, GPU, or system-level optimization background to help us improve compute performance on our car and in simulation. Experience with compiler technology is a plus. You’ll work across the entire software stack to optimize latency and compute utilization. This role follows a hybrid work schedule and reports to the TLM of the HAL Optimization team.

Requirements

  • B.Sc in Computer Science, Mathematics or a related field
  • 4+ years of industry experience
  • Strong C++ programming skills
  • Experience with CPU optimization
  • Experience with system-level optimization

Nice To Haves

  • M.Sc or PhD in Computer Science, Mathematics or a related field
  • Experience with GPU optimization (CUDA)
  • Experience with compiler technology

Responsibilities

  • Implement highly efficient sensor data processing algorithms.
  • Optimize existing CPU code.
  • Write CUDA kernels to speed up specific operations.
  • Optimize end-to-end system latency.
  • Collaborate with ML practitioners to understand their input-processing needs.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service