About the team We focus on developing high-performance GPU kernels and custom libraries that power state-of-the-art ML models’ on-vehicle inference. Our charter is to make core AI workloads faster, more reliable, and easier to maintain and deploy. That includes building custom operators when vendor libraries fall short, integrating those kernels into our ML runtime stack, and consulting on performance and CUDA debugging across the AV software stack. We collaborate closely with AI Solutions, AI Compilers, AI Architecture, and AI Tooling to ensure models can be deployed efficiently to the car while meeting strict latency and reliability targets. About the role As an AI Kernels intern, you’ll work alongside experienced kernel, compiler, and performance engineers on real production problems—profiling GPU workloads, experimenting with new kernel implementations, and strengthening the performance and robustness of the AI stack behind GM’s next-generation autonomous and assisted driving features. You’ll design, implement, and benchmark CUDA kernels and supporting infrastructure, contributing to the GPU kernels and custom libraries that improve performance, reliability, and developer experience across our ML stack.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Intern
Education Level
Ph.D. or professional degree
Number of Employees
5,001-10,000 employees