Companies’ adoption of AI agents adds autonomous workers throughout their systems. The agents read, write, schedule, log into accounts, hop through workflows and call on a list of tools in cloud platforms to finish their tasks. These powers drive the convenience – and the risk. Agents are “really enthusiastic interns” with “very little context, a desire to do a great job, and an infinite amount of energy,” Barndoor AI CEO Oren Michels told the Cybersecurity Law Report. “Without guardrails, they will make assumptions,” and, absent precise restrictions, “they will make it up.”
Like human employees, these unpredictable AI workers handle internal company data and interact with external websites. Their adoption by 80 percent of companies has led to incidents, according to reports. At Amazon, an internal coding agent decided to dismantle and rebuild a deployment environment, forcing offline parts of the AWS cloud service for 13 hours. At another company, a basic “email and calendar” agent uploaded sensitive inbox contents for a reply to an outside sender’s calendar invite, recalled Liv Porter, head of solution engineering at Gray Swan. The agent stuffed 10 emails’ contents – including credit card numbers and an offer letter – into the invite’s agenda field, mishandling a simple instruction, she told the Cybersecurity Law Report.
Beyond AI agents’ unpredictability, they introduce other threats distinct from those arising with traditional software. Agents use a new communication protocol (the Model Context Protocol, or MCP) that hackers may exploit and organizations’ engineers must learn to assess for security. Attackers could try to hijack agents’ activities or poison their memory. In multiple ways, agents expand the attack surface that organizations must defend.
This article, the second in a two-part series on real-world security for AI agents, provides an action plan for CISOs and lawyers to strengthen security and reduce risks around agentic AI, with expert perspectives from agent security specialists at Barndoor, Gravitee, Gray Swan, Skyflow and ZwillGen. It lays out practical steps and considerations for inventorying and tracking agents, protecting data, implementing human oversight and adjusting other cybersecurity controls for agents. Part one discussed three survey-based studies that showed widespread adoption of agents, immature controls, frequent incidents, and limited monitoring and tracking by companies.
See “Restricting Super Users and Zombie IDs to Increase Cloud Security” (Jul. 31, 2024).
Newfangled Threats, New Agenda for CISOs
AI agents pose several security challenges that CISOs must address – some are novel and unfamiliar, and others are variations on existing cybersecurity tasks. Security and legal teams have good justification for experiencing “fear, uncertainty and doubt” over agent threats throughout 2026. Informed explanations and analysis of agent security issues have only begun to spread to CISOs, while security tool vendors are advertising that companies need a new security stack for AI agents.
For CISOs and GCs to effectively set an agenda for securing and governing agentic AI, their teams should grasp upfront three key changes that agents bring along with their promise of automated productivity: the MCP, continuous monitoring and cyber tool adjustments.
See our two-part series on securing emerging technologies without hampering innovation: “Private Sector Challenges” (Mar. 9, 2022), and “Government Initiatives and How Companies Can Adapt” (Mar. 16, 2022).
MCP Servers, the New Communications Protocol
Anthropic created the MCP in late 2024 to create a standardized method for AI agents to access tools and other external locations. Agents build their operational power by accessing the world of tools for executing tasks, e.g., Salesforce for customer relationship management, Asana for project oversight and HubSpot for marketing. In completing tasks, agents’ reach likely will extend to many application programming interface gateways (APIs), plugin features, databases, cloud resources and messaging services, lots of which are already difficult to secure.
The development of a new standard mechanism for AI agents’ interactions, after decades of web protocols enabling communication between parties, has helped engineers embrace AI agents and greatly accelerated enterprises’ adoption of them. The AI Agentic Foundation now runs the MCP’s development as an open-source project supported by several large technology companies.
The MCP standard makes tool discovery and use easy for agents. An MCP server is a control layer wrapping an API that sets which elements of the API the AI model can access. It addresses the problem that AI agents and models lack software features to safely interact with APIs. Companies “just point the agent at a server and it is supposed to be able to read and understand which tools are available,” explained Jey Kumarasamy, a legal director in the AI division at ZwillGen.
One risk with MCPs is that some are unsanctioned, Michels noted. “There are a gazillion MCPs out there, some are available on GitHub for you to download. Most of them have malware” that might hijack the agent, he cautioned. “A Salesforce MCP might have 40 or 50 different tools that the agent can use” because “the Salesforce API is very, very powerful,” offering users many capabilities, he highlighted. Others are read-only.
Another risk is that many MCPs use a default posture of universal access, “here are all the tools,” not “here are these few tools, this way, with this data,” Michels said.
A structural pitfall with MCPs, Kumarasamy warned, is that “the servers could change at any time, and when it does change, the agent will automatically start using the new tools that are available before” the company is aware, which creates a persistent trust and diligence concern.
See “FTC Settlement Spotlights Security of APIs Proliferating Across the Internet” (Mar. 5, 2025).
Runtime Monitoring
Because of agents’ inherent unpredictability and structural trust issues like those with MCP servers, a top goal should be upgrading to continuous “runtime” monitoring to track and log agent activities. Ideally, an organization deploying dynamic autonomous AI agents would go beyond monitoring and have active controls providing alerts or circuit-breakers to stop unwanted practices by agents.
This oversight imperative is both practical and legal. To adhere to the E.U. AI Act’s record-keeping obligations under Article 12, companies may need to log for auditing all autonomous decisions in agent deployments.
While important, monitoring can be challenging. For one, some agent platforms lack native logging. Most companies probably are not yet using tools sufficient to track agents continuously. A survey by vendor Gravitee on agent-related incidents found that 70 percent of its respondents discovered incidents using a retrospective review of records, not live monitoring, the company’s chief product officer, Linus Håkansson, told the Cybersecurity Law Report.
Volume is another challenge for monitoring. AI agents, as unceasing operators, reportedly produce 10 to 20 times a human employee’s log events during active hours. Cybersecurity scanning tools on the market tout that they log every application running in a company’s systems – every file touched, command entered and network connection made. CISOs will need to research whether their endpoint detection and response service genuinely can capture AI agents’ chain of steps, from model prompts to cloud APIs to chats, documents and download sites.
See “Applying AI in Information Security” (May 15, 2024).
An Upgraded Security Stack
Do agentic systems need their own security stack? Many vendors will argue yes, as they are selling “AI” cyber tools, with terms like “AI Detection and Response.” These highlight, for example, that an agent’s detection surface goes beyond the traditional ports and processes to prompts, plans, tool calls, external links and function signatures.
Likely, some of CISOs’ typical cyber tools can help control autonomous AI. Securing agents will involve several longstanding cybersecurity measures, such as monitoring and restricted access to confidential data. Yet adjustments also will be needed to companies’ static rules to protect against agent risks.
One key area in which upgrades may be necessary is scanning AI agent outputs. “If your guardrails only detect potentially malicious actors coming in, then you’re going to miss a lot of the problems that your agent will exhibit,” Porter cautioned. Seemingly “internal” agents still may copy invites, tickets, documents or data into APIs or other external places, which data loss protection tools may not capture, she observed.
See “How to Select the Latest Cloud Security Tools and Platforms” (Aug. 21, 2024).
Action Plan for Defending and Harnessing AI Agents
GCs can help shape each company’s cyber agenda for agents by looking at liabilities. AI agents are already causing real incidents that expose the company to privacy and cybersecurity liabilities, including data leakage and unauthorized access to regulated or sensitive information. Weak visibility and identity management for agents undermine legal accountability and responses to regulators, particularly if organizations cannot fully audit agent actions or track credential use. Legal and compliance teams that lack insight into the agents’ sharing of data cannot establish proper protection for regulated data.
In the array of important steps to take to govern and secure agents set forth below, the following key themes can be found:
- Treat agents as narrowly empowered contractors/interns with their own identities and permissions.
- Keep sensitive data out of agents’ reach by default.
- Control agents with explicit and testable instructions, and shape instructions to each agent’s tool permissions and contexts.
- Monitor agents’ every action, both attempted and completed.
- Build decisional boundaries for humans to approve riskier actions.
1) Inventory and Classify
Companies should establish a single repository for information about agents. “Start by cataloging what you have – AI agents, MCP servers, LLMs, etc. – both manually and automatically, using a system that monitors and captures their use,” Håkansson recommended. Map the tools they can call, their API connections and data sources that the agents touch, he added.
Companies’ AI agent use is bound to start in decentralized fashion. “Technology has a habit of entering in the back door,” Michels observed. “People who need to get their jobs done and need to compete and want to make their bonus and want to get promoted and want to have their companies be successful are going to figure out how to use these agents” before the IT department has a chance to be involved, he said.
In an inventory, for each agent there should be the following entries:
- owners and users;
- purpose;
- identity designations;
- data it can access;
- level and duration of privileges;
- who approves privileges;
- plugins/APIs accessed;
- downstream agent interactions or integrations; and
- risk tier.
Companies should ensure central reviews of agents are collaborative and include the people who administer the various enterprise tools, Michels urged. “The CISO doesn’t know what ‘an opportunity’ is in Salesforce, let alone how to regulate who is allowed to update and change them,” he said.
Company teams can consider dividing up assessment of agents and policies concerning them. Security might approve identities, privileges, guardrails and monitoring, for example. Business or data owners might attest to the purpose, necessity and outcomes of the agents and accept the risks. Legal might approve regulated data permissions and logging policies.
See “Cybersecurity and Privacy Teams Join to Create Data Governance Councils” (May 4, 2022).
2) Establish a Policy for Each Agent
Companies should create policy IDs for each agent that articulate and encode behavioral contracts for the agent, Porter urged. Use plain‑language rules tied to the specific tools, data and agent paths, which will govern “what an agent can do. The more specific, the better,” she said.
Companies should choose more restrained tasks for agents until safeguards ramp up, which might take a year, Kumarasamy advised. For example, productive cybersecurity tasks for agents could include testing, security scans and patch proposals, he said.
Moreover, by default, agent actions and data access should be framed as “allow lists,” not blacklists, Skyflow product lead Joe McCarron recommended.
Companies also should only grant agents “read-only” rights to production systems and should not allow them to “write” or update their code, Kumarasamy advised.
Agents are excellent at calling on tools yet brittle at boundaries, so companies should block their ability to create a new account, Michels suggested. For example, he elaborated, an agent working on an employee user’s customer management tasks might reason, “You had a call with IBM. I don’t see an account for IBM. I’ll make another one.”
When setting an agent’s policy, overseers can define “never events,” like forbidding its ability to act without seeing a confirmatory email, Porter noted. “If the agent sees that it can’t behave in certain ways, in certain contexts, with certain tools, and without certain information, it will comply,” she said.
See “Strategies for Managing the Intersection of Cybersecurity and New Technologies” (Dec. 9, 2020).
3) Distinguish Agents in Identity Management
In many enterprises, agents inherit the user’s full identity and credentials. Then oversight teams decide they “want the user and the agent to have different levels of access,” McCarron told the Cybersecurity Law Report. To avoid problems, organizations should issue specific identities to agents in their systems. If an agent acts with a user’s full rights, it might bypass the company’s privilege restrictions, he noted. When the agent has its own identity, traceability improves.
Companies should consider using short-lived and tightly scoped credentials for agents, Kumarasamy suggested. They can also contain “permission sprawl” by including rotating credentials to keep an agent’s identity from lingering, and to regularly schedule the decommissioning of non-human identities. These steps can help contain incidents.
See “Checklist for Building an Identity-Centric Cybersecurity Framework” (Nov. 3, 2021).
4) Restrict Access to Data
“The number one concern that we have been hearing about from clients is the risk of data exfiltration – how do you prevent an agentic AI tool from accessing something and then uploading it to some random website,” Kumarasamy reported. “There aren’t really great ways to fix this unless one builds from the ground up very hard constraints on the agent,” he said.
To help manage the risk of exfiltration, companies should “expose as little data as possible to the AI,” McCarron urged. “Stop proliferating data, stop duplicating PII, stop spreading PII throughout your system,” he advised. Some tools can replace raw PII/PHI with non‑sensitive tokens, leaving agents with visibility only to placeholders. Using placeholders, however, might reduce agents’ utility in many cases, Kumarasamy pointed out.
Another approach to implementing data restrictions would be to use just‑in‑time approvals to use PII, but it could be difficult to implement to users’ expectations for ease and speed, Kumarasamy said. For example, McCarron shared, companies can set tools to reveal the actual sensitive values only at a moment of legitimate action, like when the agent submits a credit card.
The bottom line: sensitive data only should be revealed to agents under strict policies and ideally after considering the context.
See “Getting Used to Zero Trust? Meet Zero Copy” (Mar. 1, 2023).
5) Set Human Decision Checkpoints to Halt the Agents
Companies should evaluate where to program in human-in-the-loop gates to halt agents. High-impact actions such as financial transactions, configuration changes or data exports should require additional validation from a human.
The MCP gained an “Elicitation” feature in June 2025, which can stop the AI agent and request explicit user approval before its workflow continues, Kumarasamy highlighted as a significant development for AI governance. However, MCP creators must declare this capability to make elicitation requests to humans during the MCP’s setup, not later, so widespread use is vital.
Full user approval before actions can be unwieldy and limit agents’ usefulness, McCarron cautioned, so determining which data or actions require specific approval will be a regular discussion point for overseers.
6) Increase Logging and Observability
Companies should log agent attempts as well as actions, Michels recommended. “Safety starts with visibility,” he said, noting audit trails should show when agents tried actions, but policy or privilege blocked them. Understanding why an agent reasoned that an action was appropriate helps the security team work to adjust and ultimately harden AI security.
Companies monitoring for anomalous behavior also should see if they can program circuit breakers, rate limits or kill switches to shut down any unwanted behavior that is detected.
The hope in 2026 is that security programs will evolve to include runtime oversight tools that will evaluate agents’ intent, context and risk before they execute their actions. Currently, deployments have run a handful of agents under human supervision, but adoption of hundreds is expected before 2028, which means that “humans are not going to be able to watch this closely,” Michels cautioned. “The management of all this has to be run by agents themselves,” he said.
See “Using RegTech to Enhance Compliance” (Jun. 30, 2021).
7) Adjust Procurement for the Frenetic Market Around Agents
Many platforms are advertising that companies need tools for building, managing, securing, integrating and streamlining agents, and easing their interactions with other agents.
Amid much hype about agents, buyers at companies must beware of falling prey to “one prompt, job done.” As a tester for AI agents heading to market or into production in enterprises, Porter has replicated sensitive-data exfiltration flaws across multiple AI agentic tools. But when some startups were alerted about the discovered leakage, they did not patch their products before selling them on the market, she lamented.
See “Contracting With Vendors to Mitigate Third-Party AI Risk” (Feb. 18, 2026).
