We are building the next-generation AI infrastructure for open source and enterprise. Our work is deeply research-oriented and passionate about developing ground-breaking innovations to take state-of-the-art AI applications to the next level. What drives us is not only advancing technology, but empowering the people behind it. We are a community of researchers, engineers, and visionaries who believe that collaboration, curiosity, and openness fuel progress. If you’re motivated by impact and inspired to build tools that others can build upon, you’ll be in the right place. Mission Design and optimize large-scale pre-training systems that power Mindbeam’s generative AI models. Role Expectations • Build scalable pre-training pipelines for foundation models, optimizing throughput and efficiency. • Implement distributed training strategies across GPUs/TPUs and high-performance clusters. • Collaborate with researchers to translate experimental setups into production-ready workflows. • Develop monitoring and fault-tolerance systems to ensure reliable large-scale training. • Continuously benchmark and tune performance across hardware and software stacks. About You You thrive on scale and complexity. You enjoy solving system-level bottlenecks, pushing hardware and software to their limits, and working closely with researchers to accelerate cutting-edge AI development.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level