This job is closed
We regret to inform you that the job you were interested in has been closed. Although this specific position is no longer available, we encourage you to continue exploring other opportunities on our job board.
Tesla - Palo Alto, CA
posted about 1 month ago
This position is expected to start around May/June and continue through Aug/Sep. We ask for a minimum of 12 weeks, full-time and on-site, for most internships. In this role, you will be responsible for the internal working of the AI inference stack and compiler running neural networks in millions of Tesla vehicles and Optimus. You will collaborate closely with the AI Engineers and Hardware Engineers to understand the full inference stack and design the compiler to extract the maximum performance out of our hardware. The inference stack development is purpose-driven: deployment and analysis of production models inform the team's direction, and the team's work immediately impacts performance and the ability to deploy more and more complex models. With a cutting-edge co-designed MLIR compiler and runtime architecture, and full control of the hardware, the compiler has access to traditionally unavailable features, that can be leveraged via novel compilation approaches to generate higher performance models.