Is AI being built for the wrong work?
By Author
Alexei Dunaway
Reading Time
4
mins
Date
March 24, 2026
Share
Table of Content

Is AI being built for the wrong work?

In a recent study, researchers at Carnegie Mellon and Stanford recently mapped 43 AI agent benchmarks and 72,342 tasks against the U.S. occupational taxonomy. Their finding is worth sitting with: the AI most organizations are deploying right now was built to solve problems that represent a fraction of where real work happens.

The map most AI builders are using

The researchers used O*NET, the U.S. government's database covering all 1,016 real-world occupations, as their reference point. They measured which domains AI agents are being built and evaluated against, then compared that against where employment and wages are actually concentrated.

The mismatch is significant. AI development clusters almost entirely in computer and mathematical work, primarily programming and technical tasks. That domain represents just 7.6% of U.S. employment. Meanwhile, the work that runs most organizations barely registers.

Source: zorazrw.github.io/ai4work/

Management covers 11 million workers and over $1.3 trillion in economic value. It's 88% digital. It accounts for 1.4% of AI benchmarking effort. Legal work is 70% digitized and is nearly absent from how AI gets built and tested. Office and administrative support employs 18.2 million Americans, and it's largely invisible in the research that shapes AI development.

The researchers describe this pattern as a preference for "methodological convenience" over economic impact. It's easier to build an AI that writes code or retrieves a document than one that handles the ambiguous objectives and long-horizon dependencies of management. So that's what got built.

The skill that gets left out

What's worth noticing here is where the gap sharpens. The researchers tracked not just work domains but the underlying skills those domains require.

AI development clusters tightly around two activities: "Getting Information" and "Working with Computers." Together, those cover less than 5% of U.S. employment. The skill category with the least investment is "Interacting with Others," which the researchers describe as "practically pervading a wide range of real-world occupations."

Coaching. Coordinating. Developing people. Advising. Resolving conflict. These skills run through almost every high-value job in the economy, and they're exactly what current generic AI tools have limitation.

The data makes the picture concrete:

  • Management and Legal work are highly digitized (88% and 70% respectively) but underexplored by AI builders.
  • Agents currently struggle most with "Interacting with Others" compared to self-contained "Work Output" tasks.
  • Success rates for agents drop sharply as task complexity increases, pointing to the limits of general-purpose tools in real organizational settings.

For CHROs, this is a direct signal. The skills AI struggles with most are the skills that define effective management and people leadership. AI performs well when tasks are bounded and clearly specified. The layered, judgment-heavy, relational work that most leadership roles depend on sits largely outside what these tools were designed to handle.

Where the real opportunity sits

If we step back for a moment, the domains most underserved by AI development also carry the greatest concentration of economic value. That's not an accident. It reflects where the hard problems live, and where most vendors haven't gone yet.

The reason this gap persists is structural. Management work involves ambiguous objectives, long-horizon dependencies, and relational judgment that shifts with context. Those aren't limitations that more compute solves. They require AI built specifically for that environment. At Pinnacle, that structural gap is exactly what shaped how we built Pascal. Pascal focuses on the work that's hardest to automate and most consequential to get right: helping managers think through decisions, surface blind spots, and develop the judgment their teams depend on.

What this means for your strategy

The research points to three principles for where AI investment should head next. First, broader domain coverage beyond technical tasks. Second, realistic evaluation against complex and ambiguous scenarios. Third, purpose-built tools for human interaction rather than information retrieval.

The organizations that gain the most from AI over the next few years will be the ones that point AI at the work that actually drives organizational performance: the quality of management conversations, the development of people over time, the calibration of decisions that carry real stakes. These are areas where better performance compounds, and where the current generation of general-purpose AI tools is weakest.

The AI most commonly deployed today was shaped by what's easy to test in a research lab. The AI that creates lasting advantage is shaped by what matters most inside an organization. Those are usually very different things.

See Pascal in action.

Get a live demo of Pascal, your 24/7 AI coach inside Slack and Teams, helping teams set real goals, reflect on work, and grow more effectively.

Book a demo