Principal Research Scientist – Scaling

DatabricksSan Francisco, CA

About The Position

At Databricks, the AI Research organization is dedicated to enabling data teams to solve complex problems by building and running a leading data and AI platform. This involves developing AI models and systems using proprietary data, from pre-training LLMs to state-of-the-art retrieval-augmented generation, through novel scientific research and its production implementation. The Databricks AI Scaling team specifically focuses on advancing the efficiency of large language model (LLM) training and inference beyond current requirements. This team explores new avenues for scaling and efficiency improvements across algorithms, systems, and infrastructure. As a Principal Research Scientist – Scaling, you will lead a team of world-class researchers and engineers to push the boundaries of large-scale machine learning, with a particular emphasis on post-training, reinforcement learning (RL), and inference efficiency, optimization, and scaling. Your role will involve defining and executing a research roadmap that enhances the Databricks AI platform and delivers tangible improvements for customers in training, serving, and adapting LLMs at scale, collaborating closely with product, data, and engineering leaders to bring cutting-edge methods into production.

Requirements

  • Proven ability to lead a research team to develop novel techniques for foundation model efficiency and related topics, with a strong track record of industry impact.
  • Deep expertise in at least one of: generative AI, LLMs, distributed ML systems, model optimization, or responsible AI, with a strong emphasis on scaling and efficiency for large-scale neural networks.
  • Hands on leadership - strong programming skills and demonstrated ability to write high-quality, efficient code in Python and PyTorch for research implementation and experimentation.
  • Demonstrated ability to translate research innovation into scalable product capabilities in partnership with product and engineering teams.
  • Excellent communication, leadership, and stakeholder management skills, with experience influencing cross-functional roadmaps and aligning research with business impact.

Nice To Haves

  • Prior work at the intersection of systems and ML, such as distributed training frameworks, compiler and kernel optimization for deep learning workloads, or memory-/compute-efficient model design.
  • Strong industry and academic network in large-scale ML, with ongoing collaborations or service (e.g., PC/area chair) at top conferences in ML and systems.
  • A strong record of research impact—such as first-author publications at top ML/systems conferences (e.g., ICLR, ICML, NeurIPS, MLSys), influential open-source contributions, or widely used deployed systems—especially in optimization or efficiency.

Responsibilities

  • Lead and grow a multidisciplinary research team focused on foundational and applied AI problems, with a particular emphasis on LLM scaling, efficiency, and systems performance.
  • Define the scaling research roadmap in alignment with Databricks’ strategic objectives, prioritizing advances in foundation model efficiency and large-scale training and inference.
  • Drive algorithmic innovations for large-scale neural network training and inference, including novel optimizers, low-precision techniques, and model adaptation methods, and guide your team in rigorous empirical validation against state-of-the-art approaches.
  • Optimize end-to-end ML systems for distributed training and RL, memory efficiency, and compute efficiency through close collaboration with core systems and platform teams, ensuring that research ideas translate into performant, reliable infrastructure.
  • Partner with product and engineering to translate research breakthroughs, especially around scaling and efficiency, into customer-impacting capabilities in the Databricks AI platform.
  • Foster a culture of scientific excellence and openness, including high-quality research practices, reproducible experimentation, and effective internal knowledge sharing across Databricks AI.
  • Represent Databricks AI research externally through top-tier publications, conference talks, and collaborations with academia and the open-source community, with a focus on optimization and efficiency for large-scale models.
  • Mentor and develop talent, providing both technical guidance (research agendas, experimentation, implementation) and career development support for research scientists and engineers.
  • Define and lead independent research programs on foundation model efficiency, covering topics such as optimizer design, low-precision training/inference, scalable model architectures, and efficient adaptation methods.
  • Oversee the design and execution of large-scale experiments, including benchmarking against state-of-the-art methods and evaluating trade-offs in quality, latency, throughput, and cost.
  • Work hands-on with your team on high-quality, efficient code in Python and PyTorch for research implementation, rapid prototyping, and integration with Databricks’ production systems.
  • Collaborate with distributed systems and infra teams to push the limits of distributed training, parallelism strategies, memory management, and hardware utilization for LLMs and other large models.
  • Establish metrics, evaluation protocols, and best practices for scaling-focused research (e.g., training efficiency, inference cost, energy usage) and drive their adoption across Databricks AI.
  • Champion responsible and robust deployment of scaling innovations, ensuring that model behavior, reliability, and safety remain first-class considerations.

Benefits

  • eligibility for annual performance bonus
  • equity

Stand Out From the Crowd

Upload your resume and get instant feedback on how well it matches this job.

Upload and Match Resume

What This Job Offers

Job Type

Full-time

Career Level

Principal

Education Level

No Education Listed

Number of Employees

5,001-10,000 employees

© 2026 Teal Labs, Inc
Privacy PolicyTerms of Service