The ML Compute Platform is part of the AI Compute Platform organization within Infrastructure Platforms. Our team owns the cloud-agnostic, reliable, and cost-efficient compute backend that powers GM AI. We’re proud to serve as the AI infrastructure platform for teams developing autonomous vehicles (L3/L4/L5), as well as other groups building AI-driven products for GM and its customers. We enable rapid innovation and feature development by optimizing for high-priority, ML-centric use cases. Our platform supports the training and deployment of state-of-the-art (SOTA) machine learning models with a focus on performance, availability, concurrency, and scalability. We’re committed to maximizing GPU utilization across platforms (B200, H100, A100, and more) while maintaining reliability and cost efficiency. We are seeking a Staff ML Engineer to help build and scale robust compute platforms for ML workflows. In this role, you’ll work closely with ML engineers and researchers to ensure efficient model training and seamless deployment into production. This is a high-impact opportunity to influence the future of AI infrastructure at GM. You will play a key role in shaping the user-facing experience of the platform, ensuring that ML practitioners can discover, schedule, and debug jobs with ease. The ideal candidate brings experience in designing distributed systems for ML, strong problem-solving skills, and a product mindset focused on platform usability and reliability.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Education Level
No Education Listed
Number of Employees
5,001-10,000 employees