Updates
Common Elements of Frontier AI Safety Policies (December 2025 Update)
9 December 2025

Shared components of AI lab commitments to evaluate and mitigate severe risks.

Read more
Review of the Anthropic Summer 2025 Pilot Sabotage Risk Report
28 October 2025

External review from METR of Anthropic's Summer 2025 Sabotage Risk Report

Read more
Summary of our gpt-oss methodology review
23 October 2025

Details on external recommendations from METR for gpt-oss Preparedness experiments and follow-up from OpenAI.

Read more
Notes on Scientific Communication at METR
12 August 2025

How we think about tradeoffs when communicating surprising or nuanced findings.

Read more
What should companies share about risks from frontier AI models?
27 June 2025

Current views on information relevant for visibility into frontier AI risk.

Read more
Response to OSTP on AI Action Plan
15 March 2025

Suggested priorities for the Office of Science and Technology Policy as it develops an AI Action Plan.

Read more
Why it’s good for AI reasoning to be legible and faithful
11 March 2025

Why legible and faithful reasoning is valuable for safely developing powerful AI

Read more
Frontier AI Safety Policies
8 February 2025

List of frontier safety policies published by AI companies, including Amazon, Anthropic, Google DeepMind, G42, Meta, Microsoft, OpenAI, and xAI.

Read more
AI models can be dangerous before public deployment
17 January 2025

Why pre-deployment testing is not an adequate framework for AI risk management

Read more
Response to Bureau of Industry and Security’s proposed AI reporting requirements
11 October 2024

Red-teaming and security suggestions regarding proposed rule by the Bureau of Industry and Security, “Establishment of Reporting Requirements for the Development of Advanced Artificial Intelligence Models and Computing Clusters.”

Read more
New Support Through The Audacious Project
9 October 2024

Funding for Canary will enable research and implementation at scale

Read more
Response to U.S. AISI Draft “Managing Misuse Risk for Dual-Use Foundation Models”
8 September 2024

Suggestions for expanded guidance on capability elicitation and robust model safeguards in the U.S. AI Safety Institute’s draft document “Managing Misuse Risk for Dual-Use Foundation Models” (NIST AI 800-1).

Read more
Response to NIST Draft Generative AI Profile
2 June 2024

Comments on NIST’s draft document “AI Risk Management Framework: Generative AI Profile.”

Read more
ML Engineers Needed for New AI R&D Evals Project
16 May 2024

METR is hiring ML engineers and researchers.

Read more
Emma Abele is METR’s new Executive Director
26 April 2024

Emma moves from President to Executive Director, Beth moves to Head of Research.

Read more
2023 Year In Review
7 February 2024

A summary of what METR accomplished in 2023 – our first full year of operation.

Read more
Bounty: Diverse hard tasks for LLM agents
16 December 2023

METR (formerly ARC Evals) is looking for (1) ideas, (2) detailed specifications, and (3) well-tested implementations for tasks to measure performance of autonomous LLM agents.

Read more
ARC Evals is now METR
4 December 2023

ARC Evals is wrapping up our incubation period at ARC, and spinning off into our own standalone nonprofit.

Read more
Responsible Scaling Policies (RSPs)
26 September 2023

We describe the basic components of Responsible Scaling Policies (RSPs) as well as why we find them promising for reducing catastrophic risks from AI.

Read more
ARC Evals is spinning out from ARC
19 September 2023

ARC Evals plans to spin out from the Alignment Research Center (ARC) in the coming months, and become its own standalone organization.

Read more
Response to RfC on AI Accountability Policy
11 June 2023

Input to NTIA’s AI Accountability Policy Request for Comment.

Read more