We are building a high-density AI datacenter campus outside Austin, Texas, beginning with approximately 7MW of NVIDIA GB300 NVL72 infrastructure and scaling to 50MW+. The initial deployment is designed around real-time inference, reasoning, and high-value AI serving workloads, with a focus on monetizing capacity in live markets rather than simply leasing powered space. This is not a traditional datacenter operations role. We are hiring the person who will make the racks make money. This leader will own the strategy and execution required to turn rack-scale GPU infrastructure into a profitable inference business: selecting the right models, runtimes, orchestration stack, routing layer, pricing strategy, customer segments, and marketplace relationships to maximize revenue, uptime, and utilization. The right candidate understands that raw compute is not the business. Monetized tokens, latency-adjusted utilization, and gross margin are the business. We need a senior operator-builder who can sit at the intersection of: AI infrastructure inference performance engineering model serving and routing marketplace monetization customer / partner integration revenue optimization You will design and run the inference platform that determines how our GB300 NVL72 racks are monetized in the real-time market. That may include direct enterprise workloads, marketplace distribution, API-based reselling, model hosting, fine-tuned/private deployments, and emerging inference channels. You should know what makes money on modern inference hardware, what does not, and why. You should be able to answer questions like: Which open-weight and commercial-compatible models should run on this hardware first? How should workloads be split between premium low-latency serving, bulk throughput, reserved capacity, and experimental capacity? Should we route through third-party marketplaces, sell directly, or do both? What software stack gives us the best performance per watt, per GPU, and per dollar of capex? How do we maximize realized revenue rather than theoretical benchmark performance? How do we scale from a 7MW launch to a repeatable 50MW AI factory operating model?
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Executive
Education Level
No Education Listed