About The Position

The Seed Multimodal Interaction and World Model team is dedicated to developing models that boast human-level multimodal understanding and interaction capabilities. The team also aspires to advance the exploration and development of multimodal assistant products.

Responsibilities

  • Design and implement reinforcement learning (RL) training systems for large-scale multimodal foundation models
  • Develop unified modeling frameworks that integrate video, audio, and language, with a focus on visual latent reasoning
  • Explore RL-based approaches to bridge understanding and generation for multimodal visual reasoning
  • Collaborate with researchers to evaluate models on tasks involving world modeling, reasoning, and instruction-conditioned generation
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service