Researcher, Frontier Cybersecurity Risks

OpenAISan Francisco, CA
2d

About The Position

Models are becoming increasingly capable—moving from tools that assist humans to agents that can plan, execute, and adapt in the real world. As we push toward AGI, cybersecurity becomes one of the most important and urgent frontiers: the same systems that can accelerate productivity can also accelerate exploitation. As a Researcher for cybersecurity risks, you will help design and implement an end-to-end mitigation stack to reduce severe cyber misuse across OpenAI’s products. This role requires strong technical depth and close cross-functional collaboration to ensure safeguards are enforceable, scalable, and effective. You’ll contribute directly to building protections that remain robust as products, model capabilities, and attacker behaviors evolve.

Requirements

  • Have a passion for AI safety and are motivated to make cutting-edge AI models safer for real-world use.
  • Bring demonstrated experience in deep learning and transformer models.
  • Are proficient with frameworks such as PyTorch or TensorFlow.
  • Possess a strong foundation in data structures, algorithms, and software engineering principles.
  • Are familiar with methods for training and fine-tuning large language models, including distillation, supervised fine-tuning, and policy optimization.
  • Excel at working collaboratively with cross-functional teams across research, security, policy, product, and engineering.
  • Have significant experience designing and deploying technical safeguards for abuse prevention, detection, and enforcement at scale.

Nice To Haves

  • Bring background knowledge in cybersecurity or adjacent fields.

Responsibilities

  • Design and implement mitigation components for model-enabled cybersecurity misuse—spanning prevention, monitoring, detection, and enforcement—under the guidance of senior technical and risk leadership.
  • Integrate safeguards across product surfaces in partnership with product and engineering teams, helping ensure protections are consistent, low-latency, and scale with usage and new model capabilities.
  • Evaluate technical trade-offs within the cybersecurity risk domain (coverage, latency, model utility, and user privacy) and propose pragmatic, testable solutions.
  • Collaborate closely with risk and threat modeling partners to align mitigation design with anticipated attacker behaviors and high-impact misuse scenarios.
  • Execute rigorous testing and red-teaming workflows, helping stress-test the mitigation stack against evolving threats (e.g., novel exploits, tool-use chains, automated attack workflows) and across different product surfaces—then iterate based on findings.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service