Software Engineer, Multimodal Post-Training

DeepMindCambridge, MA
17d$141,000 - $202,000

About The Position

At Google DeepMind, we've built a unique culture and work environment where long-term ambitious research can flourish. We are seeking a highly motivated Software Engineer to join our team and contribute to groundbreaking fundamental research and deployment in large scale pre-training. About Us Artificial Intelligence could be one of humanity’s most useful inventions. At Google DeepMind, we’re a team of scientists, engineers, machine learning experts and more, working together to advance the state of the art in artificial intelligence. We use our technologies for widespread public benefit and scientific discovery, and collaborate with others on critical challenges, ensuring safety and ethics are the highest priority. The Role We're looking for a Software Engineer with strong empirical understanding of deep learning, computer vision, and large language models as well as strong engineering skills. Bonus points for experience with distributed systems or reinforcement learning. Key responsibilities: Day to day work is centered around reinforcement learning and diving into the details of data and evaluations Conduct modelling research: Use empirical and theoretical insights to derive novel research ideas that improve Gemini models. Dive deep into specific aspects of post-training (reinforcement learning, data infrastructure) to understand and improve model dynamics. Be deeply integrated with the Multimodal Vision team to focus specifically on scaling of Multimodal models. About you: In order to set you up for success as a Software Engineer at Google DeepMind, we are looking for the following skills and experience:

Requirements

  • 2-5 years of relevant professional experience as a Software Engineering or Technical role.
  • A proven track record of large scale deep learning with hands-on experience with Python and neural network training (publications, open-source projects, relevant work experience, …).
  • An in-depth knowledge of Transformer models and LLM training dynamics.
  • Ability to communicate technical ideas effectively, e.g. through discussions, whiteboard sessions, written documentation.
  • Strong knowledge and experience of Python and/or C++.

Nice To Haves

  • Experience with Large Language Models and notably Vision Language Models.
  • Working knowledge of JAX/Pytorch or similar frameworks.
  • Experience with data analysis at scale, data visualization and data ingestion pipelines for training models.

Responsibilities

  • Day to day work is centered around reinforcement learning and diving into the details of data and evaluations
  • Conduct modelling research: Use empirical and theoretical insights to derive novel research ideas that improve Gemini models.
  • Dive deep into specific aspects of post-training (reinforcement learning, data infrastructure) to understand and improve model dynamics.
  • Be deeply integrated with the Multimodal Vision team to focus specifically on scaling of Multimodal models.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service