Join us

AI Coding Tools Underperform in Field Study with Experienced Developers

AI Coding Tools Underperform in Field Study with Experienced Developers

METR ran an randomized controlled trial  (RCT) with 16 open-source devs. They tackled real-world code tasks using Claude 3.5 and Cursor Pro. The pitch: 40% speed boost. Reality: 19% slowdown. A deep dive into 246 screen recordings laid bare friction in prompting, vetting suggestions, and merging code. That friction devoured AI’s head start.

Why it matters: Teams must pair AI rollouts with RCTs. They unveil hidden snags that torpedo promised gains.


Let's keep in touch!

Stay updated with my latest posts and news. I share insights, updates, and exclusive content.

By subscribing, you share your email with @faun and accept our Terms & Privacy. Unsubscribe anytime.

Give a Pawfive to this post!


Only registered users can post comments. Please, login or signup.

Start blogging about your favorite technologies, reach more readers and earn rewards!

Join other developers and claim your FAUN.dev account now!

Avatar

The FAUN

@faun
A worldwide community of developers and DevOps enthusiasts!
Developer Influence
3k

Influence

302k

Total Hits

1

Posts