Joan Vendrell, NeuralTrust CEO and cofounder, has 15+ years of technology leadership experience advancing enterprise-grade AI security.
The year has just started, and one pattern is already clear: The pressure to deploy enterprise AI agents is not easing. It is accelerating.
In recent conversations with large enterprises, use cases have shifted from experiments to rollouts. Teams are no longer asking whether agents can automate work; they are planning how to scale autonomous systems across operations and customer workflows. What began as a few copilots is turning into a growing fleet of systems that retrieve information, make decisions and trigger actions continuously.
This shift changes the security problem. Traditional applications run within defined boundaries. AI agents operate across systems and connected services, requiring security to be designed at the system level rather than around individual components. At enterprise scale, that flexibility expands the attack surface, making implicit trust assumptions untenable.
Gartner expects this to become a mainstream deployment model quickly, predicting that 40% of “enterprise applications will be integrated with task-specific AI agents by the end of 2026.”
Why AI Agents Are Becoming A New Attack Surface
Security leaders are already feeling the impact. In my previous article, I shared some facts from my company’s CISO survey: “72% of organizations have deployed or are actively scaling AI agents across their operations, yet only 29% have comprehensive agent-specific security controls.” That imbalance matters. When autonomous systems scale faster than oversight, incidents tend to surface before governance models have time to mature.
This helps explain the growing focus on agent-specific frameworks. OWASP published the “OWASP Top 10 for Agentic Applications for 2026,” addressing risks that emerge with autonomous systems.
Frameworks help teams name the risk. The harder challenge is operationalizing them across dozens or hundreds of agents built by different teams, connected to different systems and evolving continuously.
That is the gap I see most often. Enterprises are moving fast because the business case is real, and security leaders are raising the right concerns because the threat model is new. What many teams still lack is a short, practical playbook that turns agent risk into concrete decisions before scale makes the problem harder to control.
A Short Playbook For Protecting Enterprise Agent Deployments
The playbook below reflects patterns I see repeatedly across large enterprises preparing for 2026. It is not exhaustive, but it highlights where early decisions determine whether agent programs remain governable or drift into risk, especially when security is approached with a zero-trust mindset applied to the entire system.
1. Treat agents as identities.
Most enterprises still deploy AI agents as extensions of applications or users, rather than as actors within a broader system of data, services and controls.
Autonomous agents initiate actions, access systems and operate continuously, often without direct human oversight. When they inherit credentials or permissions from early experiments, access tends to expand beyond what reflects real usage.
Teams that scale safely assign clear ownership to each agent and define permissions independently of the application it lives in. This makes it possible to answer basic questions: Which agents exist, who owns them and what systems can they act on? Without this shift, agent sprawl becomes invisible (just like with spreadsheets).
2. Design for behavior, not just access.
Traditional controls assume that correct permissions lead to safe behavior. AI agents challenge that assumption.
Two agents with identical access can behave very differently depending on context, inputs and task chaining. Many serious failures are not caused by excessive access but by sequences of actions that were technically allowed yet operationally unsafe.
As autonomy increases, security depends less on what agents can connect to and more on how they act over time. Teams should define acceptable behavioral boundaries and monitor deviations rather than relying solely on static rules. Many OWASP agentic risks surface this way, through cumulative behavior rather than single actions.
3. Assume prompt-layer exposure is permanent.
Prompt injection is not a niche problem. It is structural.
Enterprise agents ingest untrusted inputs constantly: emails, tickets, documents, chat messages and third-party tools. An attacker does not need system access. They only need to influence what the agent reads.
This is why many early incidents involve indirect manipulation. Instructions appear legitimate in isolation but lead to data leakage or unauthorized actions once combined with existing context. Teams handling this well are designed for resilience, assuming exposure will happen and focusing on containment, validation and early detection rather than perfect prevention.
4. Centralize visibility before scaling autonomy.
Almost every large enterprise underestimates how many agents it already runs.
Teams often believe they have fewer than a dozen use cases. Once integration logs and tool connections are reviewed together, the real number is often two or three times higher. This is not negligence. It is the natural outcome of accessible technology embedded in daily workflows.
Risk emerges when autonomy scales faster than visibility. Disconnected logs make it difficult to understand how agents behave and where they touch sensitive systems. Organizations that pause to establish shared visibility gain a lasting advantage.
5. Test agents the way adversaries will.
Most teams test agents for functionality, accuracy and completion. That is necessary but insufficient. Adversaries do not test whether an agent works. They test where it bends.
The most damaging failures appear at boundaries: tool chaining, memory use and action execution. These behaviors rarely surface in standard testing, yet continuous red teaming of agent behavior remains uncommon. Organizations that treat agents as living systems test them continuously, not just at launch, and revisit assumptions as capabilities evolve.
Where Enterprise AI Agent Security Goes Next
By the end of 2026, autonomous agents will be embedded across enterprise operations like cloud services today. The challenge will not be adoption but the ability to explain, govern and trust the system as a whole without relying on implicit trust.
As with every technology wave, adoption precedes structure. Enterprises that invest early in visibility, behavioral oversight and continuous assurance can turn this phase into a lasting advantage.
Forbes Technology Council is an invitation-only community for world-class CIOs, CTOs and technology executives. Do I qualify?

Leave a comment