TwelveLabs builds frontier multimodal foundation models for video understanding. Our models are deployed across a growing set of Cloud Service Provider (CSP) and data platforms — each with different compute hardware, ML inference stacks, and runtime constraints. You'll own the model-level engineering that makes this possible. This means optimizing TwelveLabs models for scalable, reliable, and performant inference across heterogeneous environments — designing how video decode pipelines, tensor orchestration, and model components behave on different hardware and inference engines. Every new platform is a new systems design problem at the model layer. You'll also design and implement massively distributed model inference systems for multimodal inputs, working across varied ML inference stacks — from hardware accelerators (NVIDIA, Trainium, Inferentia) to inference engines (vLLM, FriendliAI) and orchestrators (Ray, Anyscale). Your work directly determines how fast, how reliably, and at what cost TwelveLabs models serve inference at scale.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Education Level
No Education Listed
Number of Employees
11-50 employees