top of page

Superhuman Tools. Human Constraints.

  • Mar 2
  • 4 min read

I've watched organizations spend months building AI infrastructure while their leadership teams still need 20 meetings to make a single high-stakes decision.

The tools are getting faster. The decisions are not.

This is the real constraint. Not model capability. Not compute power. Not even data access.

It's the governance layer where tradeoffs get surfaced, exposure gets examined, and commitments get structured.

And most organizations are not set up for it.

The Productivity Paradox Returns

In 1987, economist Robert Solow noted something odd: "You can see the computer age everywhere but in the productivity statistics."

Nearly 40 years later, we're watching the same pattern repeat.

A National Bureau of Economic Research study surveyed 6,000 CEOs, CFOs, and executives across the U.S., U.K., Germany, and Australia. The finding: while 70% of businesses were actively using AI, over 80% reported no impact on company productivity or employment.

The technology is there. The results are not.

And the gap is widening. Organizations leading in AI adoption now show performance improvements 3.8 times higher than those in the bottom half. That's up from 2.7x in previous studies.

The divide is not about access to tools. It's about decision quality at the leadership layer.

The Real Bottleneck is Upstream

I've seen this pattern repeatedly: organizations adopt AI, run pilots, generate insights, and then stall.

The problem is not technological literacy. It's leadership readiness.

Only 25% of AI initiatives deliver expected ROI. Only 16% scale enterprise-wide. And only 39% of C-suite leaders believe their teams have the forward-thinking leadership needed to harness AI effectively.

The constraint has moved upstream.

As software creation becomes faster and cheaper, execution alone is no longer a differentiator. Strategy, data integrity, and decision-making now determine whether AI accelerates growth or magnifies existing dysfunction.

In many organizations, the real bottleneck is not technology.

It's human indecision.

Governance, Not Just Implementation

Here's what separates organizations that extract value from AI from those that do not:

Senior leadership actively shapes AI governance.

Enterprises where leadership is involved in governance—not just delegating to technical teams—achieve significantly greater business value.

But most organizations take the opposite approach.

Instead of leadership calling the shots with a deliberate program, many companies crowdsource AI initiatives from the ground up. They collect projects, try to shape them into something resembling a strategy, and hope for alignment.

The result: projects that may not match enterprise priorities, are rarely executed with precision, and almost never lead to transformation.

Crowdsourcing creates impressive adoption numbers. It rarely produces meaningful outcomes.

This is a governance problem. Not a technology problem.

The Quality of Framing Will Differentiate

Access to data does not guarantee better decisions.

Poor data quality costs organizations nearly $13 million annually. More critically, it undermines the accuracy, reliability, and timeliness of decisions.

But even clean data is not enough.

Most teams are not equipped to operate in an environment where analysis is no longer the constraint. They're trained to gather information, run reports, and present findings.

They're not trained to frame problems clearly or structure tradeoffs deliberately.

The ability to frame problems correctly—not just access powerful AI tools—will separate successful organizations from the rest.

Recent research emphasizes that a strong framework based on problem framing is essential. It enables leaders to justify their decisions and recognize the relationship between abstract concepts and real-world phenomena.

This becomes more important as AI becomes more capable.

Superhuman tools will become more powerful. But the quality of framing behind them will differentiate.

The Erosion of Critical Thinking

There's another tension here.

Participants who reported higher use of AI scored worse on measures of critical thinking. Younger participants showed higher dependence on AI tools and lower thinking scores than older age groups.

The tools meant to augment decision-making may be atrophying the very skills needed to use them effectively.

This is not a reason to avoid AI. It's a reason to be deliberate about how you govern its use.

If your team is outsourcing judgment to models without structuring how decisions get made, you're building dependency instead of capability.

Pressure Without Clarity

Half of CEOs believe their job stability depends on getting AI right in 2026.

Yet 60% admit they have intentionally slowed implementation due to concerns over potential errors and malfunctions.

This is the tension: high pressure, low clarity.

Leaders feel urgency but lack the governance structure to move confidently.

So they slow down. They run more pilots. They hold more meetings.

And the gap between leaders and laggards continues to widen.

What This Requires

Organizations need more architects. People who can frame problems clearly and bring disciplined thinking to complex situations.

This is not about technical expertise. It's about decision governance.

It means:

  • Clarifying the real decision beneath surface activity

  • Surfacing tradeoffs before commitment

  • Making exposure explicit so risks are selected deliberately

  • Structuring commitment so it holds under pressure

  • Defining guardrails and reassessment triggers

This is not a checklist. It's a discipline.

And it requires leadership involvement at the decision layer, not just at the approval layer.

Diagnostic Questions

If you're navigating AI adoption inside your organization, these questions may help clarify where governance is weak:

Are your AI initiatives crowdsourced or deliberately structured? If projects are emerging from the ground up without clear alignment to enterprise priorities, you're building activity instead of outcomes.

Is senior leadership shaping governance or delegating it? If governance is handled by technical teams alone, you're missing the layer where tradeoffs and exposure get examined.

Can your team frame problems clearly before building solutions? If analysis is treated as the constraint rather than problem framing, you'll optimize for speed instead of clarity.

Are decisions structured with explicit tradeoffs and guardrails? If commitments are made without surfacing what gets harder or defining reassessment triggers, drift becomes predictable.

Is critical thinking being developed or outsourced? If your team is becoming dependent on AI for judgment rather than using it to augment structured thinking, capability is eroding.

The Work Ahead

Superhuman tools are here. They will keep improving.

But they will move at human speed, constrained by judgment, trust, and context.

The organizations that extract value from AI will not be the ones with the most advanced models or the largest datasets.

They will be the ones with disciplined decision governance at the leadership layer.

The work is not to chase AI hype. The work is to strengthen the quality and durability of decisions when tradeoffs intensify and exposure becomes real.

That's where clarity compounds. And where momentum holds.

 
 
bottom of page