METR’s study on how AI affects developer productivity
In this study, developers were 19% slower when using AI.
This post was originally published in Engineering Enablement, DX’s newsletter dedicated to sharing research and perspectives on developer productivity. Subscribe to get notified when we publish new issues.
I’m summarizing a surprising new field study that challenges the prevailing narrative around AI’s impact on developer productivity. Conducted by METR, a nonprofit research organization focused on evaluating AI capabilities, the study found that AI tools actually slowed down developers working on real-world tasks. (Read the full paper here, and METR’s blog post here.)
My summary of the paper
Most AI research to date has been conducted in highly controlled settings. While these studies yield interesting findings, they often fall short of telling us how AI performs in the real world.
To close that gap, researchers at METR ran a randomized controlled trial with 16 experienced developers. Each participant had moderate experience with AI tools and at least five years of experience working on their assigned project. Here’s how the study was conducted:
- At the start of the study, developers provided a list of real issues in their repository to work on. They also forecasted how long each task would take to complete both with AI and without AI assistance.
- The tasks were then randomly assigned to either allow or disallow the use of AI tools. When AI was allowed, developers could use any tool of their choosing. In total, participants completed 246 tasks, each averaging about two hours.
- Developers would complete tasks while recording their screens, then they would self-report the total implementation time. These self-reports were used to calculate AI’s impact on task duration, while the screen recordings were used to validate the reported times and ensure no AI tools were used when they were not permitted.
- After the study’s conclusion, developers were asked to estimate how much AI tooling affected the time it took them to complete issues.
Expected vs. actual task completion time with AI
Two key findings emerged from this study. First, developers who were allowed to use AI tools took, on average, 19% longer to complete tasks. This is a significant slowdown that runs counter to mainstream expectations and headlines about AI-driven productivity gains.
Second, developers overestimated the extent to which AI helps them. Before the study, they predicted AI would speed them up by 24%. After using AI (and actually experiencing a slowdown), they still believed it had helped, estimating a 20% improvement. This disconnect wasn’t due to poor time estimation; developers were fairly accurate in predicting how long tasks would take both with or without AI. Instead, it points to overoptimism, suggesting that developers may be influenced by industry hype or their perception of the potential of AI, assuming it’s already speeding them up even when it’s not.
Possible factors contributing to the AI-related slowdown
Given the surprising results, the researchers identified five factors—supported by both qualitative and quantitative evidence—that may have contributed to the slower task completion with AI.
1. Overoptimism about AI usefulness: AI was not required to be used on AI-allowed issues. The researchers suggest that, given developers’ overoptimistic view about AI’s usefulness, they could overuse AI even when it is slowing them down.
2. High familiarity with repositories: The study found that developers were more likely to be slowed down by AI on tasks where they already had deep expertise. Before each task, they rated their “prior task exposure” on a scale of 1-5, where 1 meant “Never done this type of task before” and 5 meant “I am an expert at this type of task”. “The data showed AI was least effective when developers had high prior exposure and didn’t need external resources like documentation. Developers confirmed that AI was more helpful on unfamiliar tasks and less useful on ones they were already familiar with.
3. Large and complex repositories: Developers reported that AI tools tended to perform worse in complex environments. Since the repositories in this study were both large and mature—averaging 10 years old with over 1 million lines of code—it’s possible that AI was less effective here than in the settings examined in other studies.
4. Low AI reliability: Developers accepted less than 44% of the generations. This relatively low acceptance rate resulted in wasted time, as developers often had to review, test, and modify code, only to reject it in the end. Even when suggestions were accepted, developers reported spending considerable time reviewing and editing the code to meet their high standards.
5. Implicit repository context: Developers often rely on undocumented, tacit knowledge when working in familiar repos. AI lacked this context, making its suggestions feel off-base. One developer compared it to “a new contributor who doesn’t yet understand the codebase.”
Importantly, the researchers note these findings don’t suggest AI tools are useless—only that current tools may fall short in complex environments where developers have deep expertise.
Final thoughts
I agree with the authors’ opening line: “Despite widespread adoption, the impact of AI tools on software development in the wild remains understudied.” There’s a clear need for more real-world research like this to help engineering leaders understand what they can realistically expect from AI.
This study exposes several weaknesses of current AI tools. AI struggles in large, complex, mature codebases. It often delivers less value to experienced developers, and it is often more helpful for onboarding, unfamiliar tasks, or when documentation is lacking. Finally, the results also point to a broader pattern: AI may have a greater impact within organizations when targeted at specific, time-saving tasks rather than used as a general-purpose assistant in day-to-day development.
Perhaps most importantly, the study underscores a common challenge we see across the industry—widespread inflated expectations surrounding AI’s usefulness. The reality is far more nuanced. To realize meaningful impact, organizations should take a structured approach to AI adoption: provide training and enablement, use data to identify which use cases are actually working, and scale the use cases that deliver real value.