About the Team Our Inference team brings OpenAI’s most capable research and technology to the world through our products. We empower consumers, enterprises and developers alike to use and access our state-of-the-art AI models, allowing them to do things that they’ve never been able to before. We focus on performant and efficient model inference, as well as accelerating research progression via model inference. About the Role We’re looking for a senior engineer to design and build the load balancer that will sit at the very front of our research inference stack - routing the world’s largest AI models with millisecond precision and bulletproof reliability. This system will serve research jobs where requests must stay “sticky” to the same model instance for hours or days and where even subtle errors can directly degrade model performance. In this role, you will: Architect and build the gateway / network load balancer that fronts all research jobs, ensuring long-lived connections remain consistent and performant. Design traffic stickiness and routing strategies that optimize for both reliability and throughput. Instrument and debug complex distributed systems — with a focus on building world-class observability and debuggability tools (distributed tracing, logging, metrics). Collaborate closely with researchers and ML engineers to understand how infrastructure decisions impact model performance and training dynamics. Own the end-to-end system lifecycle: from design and code to deploy, operate, and scale. Work in an outcome-oriented environment where everyone contributes across layers of the stack, from infra plumbing to performance tuning.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Senior
Education Level
No Education Listed
Number of Employees
1,001-5,000 employees