The Red Hat AI Customer Adoption and Innovation (CAI) team is looking for a North-America based Forward Deployed AI Engineer to join our rapidly growing AI Business Unit. As inference technologies become more mainstream, our customers are seeking deep expertise in optimization, scalability, and production readiness. In this role, you will act as a bridge between engineering and the customer's environment. You will be deployed to support lighthouse customer engagements, ensuring that Red Hat AI inference products are successfully implemented, tuned, and optimized to meet specific business requirements. We are looking for a hands-on practitioner who understands that technical implementation must serve a business constraint—whether that’s cost, latency, or throughput. You will work directly with customers to design and deploy and optimize complex AI inference solutions, while simultaneously capturing those lessons to enable our wider field teams. You must have significant experience (10+ years) as consultant or technical architect. You must also have a really good understanding of inference and inference optimization, as well as practical and relevant experience. While you will have the support of the wider CAI team to upskill on specific AI technologies, you must bring a strong consulting mindset and deep technical expertise in OpenShift or Kubernetes platform engineering as well as a deep understanding of LLMs, Generative AI, and Inference. This position can be remote, but the candidate has to be located in North America, United States preferred, and willing to travel as required, up to 20% of the time.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Senior
Education Level
No Education Listed