AI Does Not Need Better Prompts. It Needs Better Specs

Move from Ad Hoc to Spec-Driven Execution

Intent Still Lives in People’s Heads

You face a system where AI optimizes locally, but the organization fails to improve as a whole.

AI coding agents amplify what already exists in your system but they do not fix it. Most software organizations were never designed as coherent execution systems. They evolved around human adaptability, where experienced engineers bridge gaps, resolve ambiguity, and compensate for missing structure through judgment and communication. That worked because humans could continuously perform knowledge work i.e. closing gaps between what is known and what must be known to act.

AI cannot operate this way. It does not “fill gaps” through context accumulation over weeks or months. It acts on what is explicitly available at the moment of execution. When intent is incomplete, fragmented, or implicit, the agent still produces output but that output reflects the gaps in the system. What looks like productivity is often just faster generation of partially correct code.

This creates a fundamental mismatch. You are applying a system optimized for human compensation to a technology that requires explicit structure. Instead of improving system performance, AI intensifies the underlying constraint: the organization’s ability to discover and encode the knowledge required to act. As the knowledge-centric framing shows, the real bottleneck is not execution - it is knowledge discovery and its availability at the point of action.

The result is predictable. Teams optimize prompts, tools, and individual workflows, but the system itself remains unchanged. Local gains appear, but they do not translate into system-level improvement.

You are not facing an AI problem but a system design problem.

AI Speed Gains Turn Into Rework

AI productivity gains plateau because fast generation creates rework loops instead of compounding output.

At first, AI makes teams look faster. More code appears, tickets seem to fly, and leaders see early signs of acceleration. But this initial speed is deceptive when the underlying intent remains vague. In a knowledge-centric view, productivity is constrained by the gap between what the team knows and what it still needs to know to complete the task, and that gap directly affects productivity.

When intent is not externalized, AI generates output before the organization has resolved enough of that missing knowledge. The result is not clean leverage but a rework loop: generate, review, correct, regenerate. Senior engineers become the absorbing mechanism for ambiguity. What looked like acceleration turns into churn, because the true constraint was never typing or code emission in the first place; it was the discovery of the knowledge needed to act correctly. The biggest impediment to productivity is lack of knowledge, and the process of knowledge discovery is the real constraint.

This is why gains flatten instead of compounding. Teams improve local output, but the system as a whole does not improve. In a systems view, knowledge is not just a property of one developer but of the organization and its accessible resources, practices, and shared artifacts. If those do not improve, faster code generation simply pushes more partially resolved work into the system.

Without structured intent, AI speeds up activity, but net productivity stalls.

Move from Ad Hoc to Spec-Driven Execution

You must decide whether AI in your organization will continue to guess intent or execute against it.

The solution is not better prompting. The solution is to move from ad hoc workflows to structured execution pipelines driven by specifications. In a spec-driven model, the specification stops being disposable documentation and becomes the primary artifact that carries the AS-IS state, the TO-BE state, and the behavioral rules that connect them. That gives both humans and AI the same explicit frame for action.

This is the real shift. Human teams have long relied on tacit knowledge, side conversations, and informal interpretation to bridge missing intent. AI cannot do that reliably. It needs a shared, durable, reviewable expression of intent that can be used upstream for planning, midstream for implementation, and downstream for validation. That is why BDD-style scenarios matter: they force intent into concrete, testable behavior instead of leaving it as abstraction.

You do not need to bet everything at once. But you do need to choose one of the listed operating models.

Prompt Optimization: Keep the current workflow and improve prompt quality, templates, and agent usage patterns.

  • Benefit: Fastest and cheapest path to incremental gains.
  • Risk: Leaves intent implicit, so delivery remains person-dependent and inconsistent.

Lightweight Spec Layer: Add structured specs, especially behavioral scenarios, while keeping the existing delivery model mostly intact.

  • Benefit: Improves clarity and creates a better shared reference for humans and AI.
  • Risk: Specs remain secondary artifacts, so execution still depends heavily on interpretation.

Full Spec-Driven Execution Model: Make specs the primary driver of planning, tasking, implementation, and validation.

  • Benefit: Creates a repeatable system where intent is explicit, reusable, and executable.
  • Risk: Requires discipline, new habits, and operating model change across roles.

Frameworks like OpenSpec and Spec Kit demonstrate this shift by embedding BDD-style scenarios directly into the specification, turning intent into a structured, testable, and executable artifact.

The winning move is not to make prompts smarter. It is to make intent executable.

Keep Improvising or Start Compounding

Your choice determines whether AI remains a local efficiency tool or becomes a system-level advantage.

If you stay in a prompt-driven model, you will keep seeing scattered wins without durable organizational improvement. A few engineers will become unusually effective. A few teams will appear ahead of others. But the performance will remain tied to individual skill in reconstructing intent from fragments. That means the organization does not actually learn how to execute better. It only learns how to rely on a small number of people who can compensate for missing structure.

This is the trap of local optimization. Prompts, agent settings, and individual craftsmanship can improve isolated outputs, but they do not create a repeatable delivery system. Variability remains high. Review load stays elevated. Rework continues to consume senior attention. Over time, leadership starts to wonder why strong early AI gains did not turn into a sustained step-change in delivery economics. The answer is simple: the system never changed.

If you move toward a spec-driven model, the consequences are different. Execution becomes more disciplined up front, but also more scalable over time. Intent is externalized once and reused many times. The same behavioral definition can guide planning, code generation, review, and testing. That reduces interpretive drift and turns AI from a fast code producer into a more reliable execution partner. The organization starts compounding knowledge instead of repeatedly reconstructing it.

Act now: accept more discipline in exchange for scalable performance.
Do nothing: accept a permanent ceiling on AI productivity.

The real decision is whether your organization will engineer intent or keep improvising it.

Next Step

Decide now to pilot a spec-driven execution model in one product area and define the minimum spec standard, covering AS-IS, TO-BE, and BDD-style behavioral scenarios, that every AI-assisted change must follow.

Dimitar Bakardzhiev

Getting started