About The Position

DayDreamer (Foundation Models, Trust & Safety) is TikTok's foundation-model team responsible for core development and research that push the boundaries of our LLM and VLM capabilities-with native trustworthiness and safety built in from day one. Our work enables product teams and customers to adopt advanced AI smoothly and responsibly, with safety woven into the model stack rather than bolted on. We are looking for talented individuals to join us for an internship in 2026. Internships at TikTok aim to offer students industry exposure and hands-on experience. Turn your ambitions into reality as your inspiration brings infinite opportunities at TikTok. PhD internships at TikTok provide students with the opportunity to actively contribute to our products and research, and to the organization's future plans and emerging technologies. Our dynamic internship experience blends hands-on learning, enriching community-building and development events, and collaboration with industry experts. Applications will be reviewed on a rolling basis - we encourage you to apply early. Please state your availability clearly in your resume (Start date, End date).

Requirements

  • Currently pursuing an PhD in Computer Science or a related technical field.
  • Research experience in at least one of: LLMs, AI Safety, Computer Vision, Multimodality.
  • Actively track recent AI-safety developments (familiar with current papers, benchmarks, and terminology).
  • Proficient with at least one deep learning framework (e.g., PyTorch, TensorFlow).
  • Excellent analytical/problem-solving ability, clear logical thinking, and strong communication/collaboration skills.

Nice To Haves

  • Publications in top AI or Security venues (e.g., NeurIPS, ICLR, ICML, COLM, CVPR, ICCV, ECCV, USENIX Security, ACM CCS, IEEE S&P, NDSS).

Responsibilities

  • Pretraining & Continued Pretraining (CPT): Explore and develop approaches that improve general capability and safety for LLMs and VLMs via pretraining/CPT.
  • Evaluation & Measurement: Build advanced evaluation systems to study emerging LLM/VLM skills and safety behaviors, aligned with real downstream use.
  • Post-Training & RL: Develop advanced reinforcement-learning strategies (e.g., RLHF/RLAIF/DPO variants), and investigate how to balance pretraining and post-training for better capability and safety alignment.
  • Scaffolded Settings & Agents: Probe new failure modes and pitfalls in scaffolded environments (agents, workflows, tool use), and translate insights into robust mitigations.

Stand Out From the Crowd

Upload your resume and get instant feedback on how well it matches this job.

Upload and Match Resume

What This Job Offers

Career Level

Intern

Industry

Broadcasting and Content Providers

Education Level

Ph.D. or professional degree

Number of Employees

5,001-10,000 employees

© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service