AI Can Do More Than We Use It For: The Massive Gap Between Capability and Reality
Anthropic's landmark research shows AI could theoretically perform the majority of tasks in most white-collar occupations. But actual usage tells a very different story. Here's what the gap means for your career.
Series: Decoding Anthropic's AI Labor Market Research
The Research That Changes How We Talk About AI and Jobs
On March 5, 2026, Anthropic published what may be the most important piece of research on AI and employment to date: "Labor market impacts of AI: A new measure and early evidence." Unlike previous studies that relied purely on theoretical assessments, this paper introduces a new metric called observed exposure that combines what AI could do with what it is actually doing in the real world.
The findings are both reassuring and sobering. The reassuring part: AI is far from replacing most jobs today. The sobering part: the theoretical ceiling for AI disruption across white-collar work is enormous, and the gap between capability and adoption is closing every month.
This is Part 1 of our three-part deep dive. In this post, we break down the study's core framework and what the massive gap between AI's theoretical capability and observed real-world usage means for career planning.
Understanding the Radar Chart: Two Very Different Stories
The most striking visual from Anthropic's paper is a radar chart (Figure 2 in the original) that overlays two measurements across 22 major occupational categories:
Theoretical AI Coverage (Blue)
Based on the Eloundou et al. (2023) framework, this measures whether it is theoretically possible for an LLM to make a task at least twice as fast. A task scores 1.0 if an LLM alone could double its speed, 0.5 if it needs additional tools, and 0 if it's beyond AI's reach.
Observed AI Coverage (Red)
Derived from the Anthropic Economic Index—actual Claude usage data—this measures which tasks are actually being performed with AI in real-world professional settings. It weights automated use more heavily than augmentative use.
The contrast is dramatic. The blue area (theoretical capability) sprawls outward like a large, spiky polygon, covering the vast majority of tasks in knowledge-work categories. The red area (actual usage) is a small, tight cluster near the center. In every single category, observed usage is a fraction of what's theoretically possible.
The Numbers That Matter: Category by Category
Let's walk through what the radar chart reveals for the most important occupational categories:
Computer & Math
Gap: 61ppThe highest observed coverage of any category, yet still only a third of what's theoretically possible. This includes programmers, data analysts, and systems administrators.
Management
Gap: ~70ppNearly all management tasks are theoretically feasible for AI, but adoption remains low. Strategy, planning, and coordination tasks show limited real-world AI penetration.
Office & Admin
Gap: ~70ppData entry, scheduling, and document processing are theoretically automatable, but organizational inertia keeps adoption low.
Legal
Gap: ~60ppContract review, legal research, and document drafting are feasible for AI. But regulatory constraints, liability concerns, and verification requirements slow adoption.
Business & Finance
Gap: ~55ppFinancial analysis, reporting, and forecasting are theoretically exposed. Observed usage is growing but still limited.
Arts & Media
Gap: ~53ppContent creation, editing, and design tasks are feasible. But creative judgment and brand voice keep human involvement high.
Construction
Gap: ~12ppPhysical tasks like operating machinery and on-site work remain firmly beyond AI's reach. The small theoretical coverage relates to planning and documentation.
Why the Gap Exists: Five Barriers to AI Adoption
If AI can theoretically handle 94% of Computer & Math tasks, why is actual usage only at 33%? The research identifies several key friction points:
Model Limitations
Current LLMs aren't quite good enough for many tasks that are theoretically feasible. Quality, reliability, and hallucination issues limit deployment.
Legal & Regulatory Constraints
Tasks like authorizing drug refills or representing clients in court face legal barriers even if AI is technically capable.
Software Integration Requirements
Many tasks need AI embedded into specific enterprise software, not just a chatbot interface. This integration takes time.
Human Verification Steps
High-stakes outputs (financial reports, medical recommendations, legal opinions) require human review, slowing full automation.
Organizational Inertia
Companies are slow to restructure workflows, retrain staff, and update processes even when AI tools are available.
The Closing Gap: Why This Matters for Career Planning
Here's the critical insight: the gap between blue and red is not static. It's closing. As AI models improve, as enterprise software integrates LLMs, as regulations adapt, and as organizations experiment with automation, the red area on that radar chart will expand outward toward the blue.
The research validates this trajectory: 97% of tasks actually observed in Claude usage fall into categories rated as theoretically feasible. This means real-world usage is tracking theoretical capability almost perfectly—it's just not there yet in volume.
Think of it this way: the blue area is a map of where AI will go. The red area is where it already is. The space between is your window of opportunity.
What This Means for You
- If you're in a high-theoretical, low-observed category (Management, Office & Admin, Legal): You have time, but the wave is coming. Start learning AI tools now while adoption is still low and early movers have an advantage.
- If you're in a high-theoretical, high-observed category (Computer & Math, Business & Finance): The transformation is already underway. AI proficiency isn't optional—it's the new baseline for competitiveness in your field.
- If you're in a low-theoretical category (Construction, Healthcare Support, Protective Services): Your job's core tasks remain beyond AI's reach. But even here, AI fluency for adjacent tasks (documentation, planning, communication) will become valuable.
A New Way to Measure AI Risk: "Observed Exposure"
Previous attempts to measure which jobs AI would affect relied on theoretical assessments—experts guessing what AI could do. The problem? Past predictions about technology and jobs have a poor track record. The paper notes that a prominent study on job offshorability identified a quarter of US jobs as vulnerable, but a decade later, most maintained healthy employment growth.
Anthropic's "observed exposure" metric is different because it combines three data sources:
- O*NET database: The US government's catalog of tasks for ~800 occupations
- Eloundou et al. (2023) theoretical ratings: Whether each task could theoretically be doubled in speed by an LLM
- Anthropic Economic Index: Actual Claude usage data showing which tasks are being performed with AI in practice
The measure weights automated use (API integrations, fully automated workflows) more heavily than augmentative use (humans using AI as a helper). This is important: a customer service chatbot handling tickets autonomously signals more disruption potential than a lawyer using Claude to brainstorm case strategy.
The Historical Lesson: Predictions Are Hard
The researchers are refreshingly humble about prediction. They point out that:
- Studies on the employment effects of industrial robots reach opposing conclusions
- The scale of job losses from the China trade shock continues to be debated
- The government's own occupational growth forecasts have added little predictive value beyond simple trend extrapolation
This is precisely why the "observed exposure" approach matters. Instead of guessing the future, it measures the present and tracks how the gap closes over time. It's a thermometer, not a crystal ball.
The Bottom Line
The radar chart tells a story of two realities:
The Theoretical Reality
AI could theoretically accelerate the majority of tasks in most knowledge-work categories. For Computer & Math (94%), Office & Admin (90%), Management (~95%), and Legal (~85%), the ceiling is near-total.
The Observed Reality
Actual AI usage covers a fraction of what's possible. Even the most penetrated category (Computer & Math) is only at 33%. Most categories are below 25%. The AI revolution is still in early innings.
For career planning, this gap is both a warning and a window. The theoretical coverage tells you where disruption will eventually reach. The observed coverage tells you how much time you have to prepare. The smart move is to use that time wisely.
In Part 2, we dive into which specific occupations are most exposed and what the demographics of at-risk workers actually look like. The findings may surprise you.
How Exposed Is Your Career?
Get a personalized AI exposure analysis for your specific role and industry. Our assessment uses real data to help you understand your risk and build a plan.
Take the Assessment