At Fireworks, we're building the future of generative AI infrastructure. Our platform delivers the highest-quality models with the fastest and most scalable inference in the industry. We've been independently benchmarked as the leader in LLM inference speed and are driving cutting-edge innovation through projects like our own function calling and multimodal models. Fireworks is a Series C company valued at $4 billion and backed by top investors including Benchmark, Sequoia, Lightspeed, Index, and Evantic. We're an ambitious, collaborative team of builders, founded by veterans of Meta PyTorch and Google Vertex AI. Fireworks is hiring a deeply technical, systems-minded PM who is obsessed with building the most performant, reliable, and scalable GPU inference platform in the world. You'll work on product strategy for Fireworks' core infrastructure-spanning multi-region deployments, autoscaling experiences, new inference verticals, and user-facing packaging. You will partner directly with customers to deeply understand their workload patterns, reliability needs, and performance requirements. You will convert these insights into an infrastructure roadmap with Fireworks' core engineering and field team.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Industry
Publishing Industries
Education Level
No Education Listed
Number of Employees
101-250 employees