Spun out of MIT CSAIL, we build general-purpose AI systems that run efficiently across deployment targets, from data center accelerators to on-device hardware, ensuring low latency, minimal memory usage, privacy, and reliability. We partner with enterprises across consumer electronics, automotive, life sciences, and financial services. We are scaling rapidly and need exceptional people to help us get there. Our inference stack is central to everything we ship. You'll be a core part of the team responsible for the engine layer that runs our models in production and in partner environments, and for the benchmarking infrastructure we use to evaluate our own work and verify what partners bring to us. Day to day, that means working closely with research and product, but also directly with external engineering teams. We need someone who: Can pick up unfamiliar tools quickly and knows how to assess whether they're worth using. Designs AI benchmarks and holds methodology to a high standard. Cares about inference details, understands the tradeoffs, and checks what changed across the board before calling something done. Doesn’t consider a model port finished until you can prove the outputs are correct.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Education Level
No Education Listed