In this role, you will be responsible for driving down wall-clock time to convergence by profiling and eliminating bottlenecks across the foundation model training stack, from data pipelines to GPU kernels. You will design, build, and optimize distributed training systems using PyTorch for multi-node GPU clusters, ensuring scalability, robustness, and high utilization. Additionally, you will implement efficient low-level code using CUDA, cuDNN, Triton, and custom kernels, integrating it seamlessly into high-level training frameworks. Your work will also involve optimizing workloads for hardware efficiency, focusing on CPU/GPU compute balance, memory management, data throughput, and networking. Furthermore, you will develop monitoring and debugging tools for large-scale runs, enabling rapid diagnosis of performance regressions and failures.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Career Level
Senior