Planckpoint Logo
Why Knowing What You Want Is Harder Than Getting It Done, Especially in Artificial Intelligence Efforts

Artificial Intelligence

Why Knowing What You Want Is Harder Than Getting It Done, Especially in Artificial Intelligence Efforts

The central challenge for leaders today isn't just execution, which AI can handle, but defining with absolute clarity what needs to be done. This article explores why strategic clarity is the new, critical discipline in the age of AI.

By Florian Scheibmayr|January 28, 2026

"Would you tell me, please, which way I ought to go from here?"

"That depends a good deal on where you want to get to," said the Cat.

"I don't much care where,” said Alice.

"Then it doesn't matter which way you go," said the Cat.

- Lewis Carroll, Alice's Adventures in Wonderland

When Execution: The Discipline of Getting Things Done was published in 2002, it captured the dominant management challenge of the era. Most organisations had a reasonable sense of direction. Strategies were articulated, targets were set, and priorities were broadly understood. The problem was execution. Leaders struggled with follow-through, coordination, and operational discipline.

That challenge no longer defines the moment. Today, artificial intelligence executes instructions with a level of consistency, speed, and endurance that human systems never could. What was once the hardest part of management is now automated, scalable, and increasingly inexpensive. Execution, at least in its traditional sense, has become easier. The difficulty has moved upstream. The central leadership challenge today is not whether a strategy can be executed. It is whether leaders can specify, with sufficient clarity and completeness, what should be executed in the first place.

When execution stops being the bottleneck

Modern AI systems do not hesitate, reinterpret, or apply judgment. They execute exactly what they are instructed to do, continuously and at machine speed. As a result, the core managerial question has shifted. It is no longer “Can we make this happen?” but “Have we told the system precisely what should happen, and what must never happen?” When execution becomes trivial, strategic clarity becomes critical.


Why strategic clarity is now harder than execution

Traditional goal setting today is incomplete unless its constraints are made explicit. Acceptable response times, quality thresholds, escalation rules, prohibited tactics, compliance requirements, and ethical red lines must all be specified. The challenge is not goal setting. Organisations solved that decades ago. The challenge is constraint completeness. Leaders often discover their unstated assumptions only when an AI system violates them while still meeting the target. When retention improves through manipulative tactics, the metric is achieved, but the outcome is wrong.

In the early 2000s, SMART goals were sufficient because humans supplied what was left unsaid. A target such as “Increase customer retention from 78 per cent to 85 per cent by Q2” worked because employees understood the boundaries. They knew not to deceive customers, violate privacy, or undermine trust, even when those limits were not written down. Judgement closed the gaps. AI does not do that. What humans once inferred now must be defined to perfection.


Rethinking people, strategy, and operations

This shift fundamentally alters the familiar triad of people, strategy, and operations. In people processes, high performance once meant operational discipline and consistent execution. Today, the most valuable capability is judgment. Organisations must reward those who can identify when a system is executing the wrong objective efficiently. Ethical reasoning, systems thinking, and governance literacy matter more than speed or output. In strategy, direction is no longer enough. Leaders must translate intent into executable specifications that include constraints, trade-offs, escalation triggers, and verification mechanisms. Direction without specification invites failure at scale. In operations, the focus has moved from doing to governing. Operations now involve monitoring automated execution at granular levels, detecting drift, intervening when context changes, and enforcing oversight even as systems operate faster than human decision cycles. The mismatch between machine speed and human governance introduces risks that did not exist in earlier eras.


What has not changed

Despite these shifts, a core insight from 2002 still holds. Execution cannot be delegated. What has changed is the form it takes. Leadership now means owning strategic clarity with the same intensity once applied to follow-through. It requires understanding the limits of AI, building governance-focused cultures, and maintaining the courage to override automated recommendations when they conflict with values or long-term outcomes. Uncertainty has not disappeared. It has become more compressed and more consequential.


Getting somewhere is no longer the problem

It is easy to forget that getting somewhere is not the same as getting to the right place. With AI, organisations can move very quickly. That is precisely the risk.

AI solves the execution problem that dominated management thinking in the early 2000s, but it creates a deeper challenge. Humans are no longer naturally positioned to govern systems that act faster than judgment can keep pace. Organisations that succeed in the AI era will master strategic clarity as rigorously as earlier leaders mastered execution discipline. They will articulate objectives with precision, define boundaries with care, and preserve human judgement as automation scales.


The discipline of this decade is not getting things done. It is ensuring that where you end up is where you intended to go.