We are seeking a Researcher in Privacy-Preserving Safety to help design and build the next generation of privacy-preserving safety systems for frontier AI models. This role sits at the intersection of AI safety, security, and privacy, with a focus on developing auditable, privacy-first mechanisms that enable robust harm detection and mitigation without exposing sensitive user data. You will help define and operationalize frameworks for identifying and addressing frontier risks (e.g., bioweapon instructions, malware creation, suicide/self-harm risks, jailbreaks), while ensuring that privacy guarantees remain intact—even under adversarial conditions. This role is central to our long-term goal of scaling our automated privacy-preserving safety systems to mitigate potential harms while minimizing human review. You’ll work on foundational problems such as privacy-preserving monitoring, algorithmic auditing, secure enclaves, and adversarially robust safety enforcement protocols, helping ensure that safety systems scale without compromising user trust.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Education Level
Ph.D. or professional degree
Number of Employees
1-10 employees