We are a nonprofit research organization that develops scientific methods to assess AI capabilities, risks and mitigations, with a specific focus on threats related to autonomy, AI R&D automation, and alignment. We believe it is robustly good for civilization to have a clearer understanding of what dangers AI systems pose, and we are extremely excited to find ambitious, excellent people to join our team and tackle one of the most important challenges of our time. We evaluate candidates primarily through work tests. We usually do an in-person trial as well but can be flexible about this. METR currently has 3 primary research streams: Capabilities: Accurately measuring frontier model performance on threat-relevant tasks (autonomy, AI R&D automation, etc.) and predicting future capabilities. We develop and maintain benchmarks, diverse evidence-gathering methods, and metrics to track capability trends and anticipate the thresholds that matter most for safety. Monitorability: Understanding how well frontier models can take subversive or unwanted actions despite various monitoring or control protocols. We build the research infrastructure â novel metrics, control evaluations, elicitation methods â needed to improve the world's understanding of how effectively current and future models can circumvent oversight. Alignment/Propensity: Determining whether or not a model that is capable of causing catastrophic harm (in its actual deployment setting) would be likely to actually do so in a given high-stakes deployment setting. We aim to develop the science of propensity evaluations and examine when we might expect high-stakes catastrophic misalignment. The Capabilities and Monitorability streams are both hiring Research ICs (individual contributors), while the Alignment/Propensity stream is hiring for a Research Stream Lead, followed by Research ICs down the line. The stream you end up joining will be based on a combination of working fit and interest. For our Research IC roles, we are looking for a combination of skills across âresearch scienceâ, âresearch executionâ and software engineering. You may not have all of these skills (for example, we donât expect software engineering to be a large part of the role for narrowly focused researchers). For the Research Stream Lead role, we are additionally looking for research management skills. We're seeking a researcher to help us better understand AI capabilities. Previous work in this vein includes agent time horizons, a commonly-used metric for measuring AI progress, and RCTs on open-source developer productivity.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Education Level
No Education Listed