Vivek Acharya, AI Strategist | Author | AI Ethics Assessor | Advancing AI Agents, Process Automation and Responsible AI.
By month 10 of an AI pilot that’s “almost ready,” the pattern is familiar: The demo keeps improving, the slide deck evolves, but the real impact stays vague. These are zombie pilots, projects not alive enough to scale, yet not dead enough to stop consuming time, budget and credibility.
Agentic AI (systems that can plan and take actions across tools) makes the zombie problem more common, not less. In fact, Gartner predicts that over 40% of agentic AI projects will be canceled by the end of 2027 due to escalating costs, unclear business value or inadequate risk controls. But that’s not a reason to avoid AI agents altogether. It’s a reason to pilot them with discipline.
Why Zombie Pilots Happen
Zombie pilots don’t linger because teams are lazy. They linger because “success” was never defined for the messy middle, the phase where edge cases show up, permissions get complicated and the workflow stops looking like the polished demo. Three forces often keep these pilots undead:
• Demo Bias: Curated examples hide operational noise.
• Sunk Cost: Stopping feels like admitting defeat after public sponsorship.
• Split Ownership: Business, IT, security, data and legal all touch the workflow, so decisions stall.
Kill Criteria Is An Operating Discipline
Strong programs decide upfront what would make them stop a pilot. Not because they expect failure, but because they want speed without self-deception. A cadence that works: week two, week six and week 12. At each gate, leadership makes one call, scale, pivot or stop, based on evidence.
A Kill-Criteria Scorecard Leaders Can Use
Keep the scorecard to one page with five criteria, supported only by factual evidence. Here are five key criteria and their “kill signals”:
1. Outcome Clarity (Measurable And Owned): Choose one primary metric (cycle time, cost per case, error rate, risk reduction), set a baseline, and name one accountable owner. Kill signal: “better/faster/smarter” with no baseline and no owner.
2. Data Readiness (No “We’ll Fix It Later”): Confirm the system of record, test data completeness on real cases and define what data is off-limits. Kill signal: The pilot depends on manual cleanup, special extracts or “temporary” access.
3. Exception Rate (The truth Serum): Track how often humans override, rewrite or escalate the agent’s work, and separate tunable issues from structural ones. Kill signal: Override rates stay high after initial tuning, or the top drivers are structural.
4. Integration Friction (Tools And Permissions): List every system the agent must touch, the minimum privileges required and the time-to-access in your environment. Kill signal: Integration effort or access risk outweighs the value.
5. Governance Feasibility (Auditability, Rollback, Escalation): If an agent can act, leaders will ask: Can we trace what happened, and can we undo it? Minimum bar: action logging, rollback, escalation. Kill signal: You can’t meet baseline traceability, rollback and human-override expectations for the domain.
How To Use The Scorecard Without Killing Momentum
Treat the scorecard like a release checklist. At each gate (week two/six/12), require evidence: early movement on the outcome metric, real-case samples (including failures), an integration plan with owners and an agreed risk/control posture. If one area is yellow, pick one fix and re-test quickly. If it’s red on outcome, data readiness or governance, stop and write a one-page learning memo so the next pilot starts smarter.
The Two Doors Rule
Before you pilot an agent, decide which door it sits behind:
• Door A (Low Risk): The worst case is inconvenience. The agent drafts, summarizes or suggests routing, while a human approves before anything is executed.
• Door B (High Stakes): Money movement, access rights, regulated decisions, customer entitlements or anything with safety/compliance implications.
If the team can’t name the door, that uncertainty is your first finding.
When A Great Demo Turns Into A Zombie
I led an agentic pilot in insurance operations aimed at a practical workflow: Read inbound policyholder and agent emails, pull context from internal systems (policy and claims history) and draft a response for a service rep to approve.
In demos, it looked great because the cases were clean. In production-like traffic, emails arrived without policy numbers, IDs didn’t reconcile between systems and attachments were scanned and inconsistent. The agent still produced drafts, but humans spent more time checking and correcting than they would have spent writing from scratch.
By week six, the human override rate was about 70%. That single metric changed the conversation. The question wasn’t “Is the model smart?” It was “Are we scaling rework?”
We hit an uncomfortable tradeoff: speed versus control. The business wanted a broader rollout to learn faster; risk and security wanted tighter guardrails before any autonomy touched sensitive workflows. We pivoted into Door A: The agent stayed in draft mode, with read-only access, focused on the most common low-risk request types. And we set a hard rule: If required fields were missing or confidence was low, the agent routed the case to a human instead of guessing.
It wasn’t glamorous, but it prevented a multi-year zombie pilot and created a foundation we could trust.
Why Kill Criteria Accelerates The Winners
Having clear kill criteria actually speeds up and makes innovation safer. Finance teams stop seeing AI initiatives as endless science experiments, because there’s a clear decision process and budget guardrails. Risk and compliance teams feel involved rather than bypassed, because governance checks are baked in. Frontline employees stop feeling like perpetual beta testers for every new AI toy, because only the pilots that prove real value get to graduate. And when you do get a winner, scaling it is much quicker because you’ve already proven the outcomes, integration hooks and control mechanisms on a small scale.
If you want agentic AI to become more than a buzzword in your organization, start every pilot by defining the conditions under which you’ll kill it. Then have the discipline to honor that decision when the gate arrives. Ironically, knowing when you would stop actually gives you the freedom to move faster, with confidence, on the initiatives that deserve to live.
Forbes Technology Council is an invitation-only community for world-class CIOs, CTOs and technology executives. Do I qualify?

Leave a comment