Original: Swyx · 19/02/2026
Summary
Most Claude Code turns are short. The median turn lasts around 45 seconds. There’s a lot of small tidbits going on, with former guest [Fei-Fei Li’s World Labs](https://finance.yahoo.com/news/ai-pioneer-fei-fei-lis-192214332.html?guccounter=1&guce_referrer=aHR0cHM6Ly9rYWdpLmNvbS8&guce_referrer_sig=AQAAAIHn6aL6ECAJH2dSErr8YVZLehWRdwRA_q2KzFp8_WzVMfX6CWRlOPG8iQwhJU7OBw8yR61sFKey Insights
“Most Claude Code turns are short. The median turn lasts around 45 seconds.” — Discussing the typical duration of interactions with Claude Code.
“The METR evaluation captures what a model is capable of in an idealized setting with no human interaction and no real-world consequences.” — Comparing METR’s theoretical model capabilities with Anthropic’s practical findings.
“new users start off with 20% auto-approve, and increase to >50% over time with experience.” — Highlighting how user interaction with Claude Code evolves with experience.
Topics
Full Article
Published: 2026-02-19
Source: https://www.latent.space/p/ainews-anthropics-agent-autonomy
There’s a lot of small tidbits going on, with former guest Fei-Fei Li’s World Labs and The Era of Experience’s David Silver both raising monster $1B rounds, and Anthropic officially blocking OpenClaw using Claude OAuth tokens (consistent with post-OpenCode policy), with OpenAI employees politely reminding everyone that they’re more than welcome to use OpenAI plans instead on the same day (complete coincidence, we are sure). However, all that will pass. What we’d highlight today is Anthropic’s study of its own API usage patterns, Measuring AI agent autonomy in practice. As you might expect, most usage is coding, but you can start to go down the list of the other uses and pick off the next likely targets for agents:
The METR evaluation captures what a model is capable of in an idealized setting with no human interaction and no real-world consequences. Our measurements capture what happens in practice, where Claude pauses to ask for feedback and users interrupt. And METR’s five-hour figure measures task difficulty—how long the task would take a human—not how long the model actually runs. (also… Most Claude Code turns are short. The median turn lasts around 45 seconds, and this duration has fluctuated only slightly over the past few months (between 40 and 55 seconds). In fact, nearly every percentile below the 99th has remained relatively stable.Enough said. Because Anthropic has full access to Claude Code telemetry, there are other autonomy measures nobody else has. For example… new users start off with 20% auto-approve, and increase to >50% over time with experience
Agentic coding + harness engineering (Claude Code, Cursor, LangSmith, Deep Agents, SWE-bench process)
Key Takeaways
Notable Quotes
Most Claude Code turns are short. The median turn lasts around 45 seconds.Context: Discussing the typical duration of interactions with Claude Code.
The METR evaluation captures what a model is capable of in an idealized setting with no human interaction and no real-world consequences.Context: Comparing METR’s theoretical model capabilities with Anthropic’s practical findings.
new users start off with 20% auto-approve, and increase to >50% over time with experience.Context: Highlighting how user interaction with Claude Code evolves with experience.
Related Topics
- [[topics/claude-code]]
- [[topics/ai-agents]]
- [[topics/anthropic-api]]
Related Articles
[AINews] OpenAI and Anthropic go to war: Claude Opus 4.6 vs GPT 5.3 Codex
Swyx · explanation · 85% similar
[AINews] "Sci-Fi with a touch of Madness"
Swyx · explanation · 81% similar
Effective harnesses for long-running agents
Anthropic Engineering · how-to · 80% similar
Originally published at https://www.latent.space/p/ainews-anthropics-agent-autonomy.