Anthropic’s new measurement work tracks millions of human‑agent sessions to see how autonomy evolves when the tools leave the lab. The headline: operators tend to grant more unattended time as they gain experience, suggesting growing trust in constrained settings. Crucially, this is not blind trust, since usage still concentrates on reversible actions. The data points to a pragmatic path where ambition expands inside safety rails. 🔍🧩 anthropic.com
The nuance matters. Experienced users are more likely to auto‑approve actions, yet they also interrupt more often, which hints at a sharper, more surgical oversight style. Most actions remain low risk, so interruption costs stay manageable. This aligns with the “exoskeleton” framing that tools should amplify humans on discrete tasks while judgment stays human‑led. The takeaway is augmentation first, automation second. 🧠🛡️ anthropic.comkasava.dev
Zooming out, the study’s call for stronger post‑deployment monitoring and new interaction paradigms pairs with two adjacent currents. One is the push for baseline literacy and guardrails to prevent a rising tide of low‑quality code from non‑experts. The other is fresh capital for independent alignment research that can harden evaluation and governance. Together they sketch an ecosystem where capability, operator skill, and safety infrastructure mature in lockstep. 🚦📈 anthropic.comvibingwithai.su...openai.com