The AI Control Problem

AI Governance IS Cybersecurity

TV
Thiago Victorino
11 min read
AI Governance IS Cybersecurity

In November 2025, Anthropic published a technical report documenting something that had never been publicly confirmed before: an AI system orchestrating a cyber espionage campaign from reconnaissance through data exfiltration. The operation, cataloged as GTG-1002, showed an AI performing 80 to 90 percent of the attack chain. A human set the objective. The machine did the rest.

Anthropic has a dual interest in disclosing this. Their model was misused, and the responsible disclosure generates trust. Both things can be true simultaneously. What matters is what the incident reveals about organizational structure: the team responsible for AI governance and the team responsible for cybersecurity are looking at the same threat from different org chart positions. One team sees an AI risk. The other sees a cyber risk. Neither has full authority to respond.

That structural separation is the vulnerability.

The Convergence Nobody Planned

In The Governance Layer, we examined what the February 2026 cybersecurity selloff revealed: detection is being commoditized while the governance surface expands. That article tracked the market signal. This one tracks the operational consequence.

The operational consequence is that AI governance and cybersecurity are collapsing into a single discipline, whether organizations want them to or not. Three independent developments are forcing this convergence.

First, NIST published its Cyber AI Profile (NISTIR 8596) in draft form in December 2025. The document merges the Cybersecurity Framework 2.0 with the AI Risk Management Framework. Call it what it is: the first formal acknowledgment by a standards body that these two domains cannot be governed separately. The controls overlap. The risks overlap. The response procedures overlap. Maintaining two separate frameworks for a single threat surface is an organizational choice, not a technical requirement.

Second, McKinsey published research in October 2025 identifying five risk categories unique to agentic AI: chained vulnerabilities across multi-agent systems, cross-agent task escalation, synthetic-identity risk, untraceable data leakage, and data corruption propagation. McKinsey, which also provides cybersecurity consulting services, framed AI agents as “digital insiders.” The analogy is precise. An AI agent with system privileges behaves like a senior employee with broad access, except it operates at machine speed, does not take breaks, and does not notice when its instructions have been subtly altered through prompt injection.

Third, the economics have already settled the argument. IBM’s 2024 Cost of a Data Breach Report, based on 604 real breaches across its 19th annual cycle, found that organizations with extensive AI and automation in their security operations averaged $3.84 million per breach. Organizations without: $5.72 million. The difference is $1.88 million per incident.

That $1.88 million represents the savings from having governance structures that allow AI to function within security operations, not the cost of buying tools. The tools without the governance produce the $5.72 million outcome.

The Digital Insider Problem

McKinsey’s “digital insider” framing deserves serious attention because it reframes a governance problem that most organizations have misclassified.

Identity and access management was designed for humans. Humans have predictable patterns. They log in from known locations. They access systems during working hours. They escalate privileges through request chains. When a human behaves anomalously, security systems flag it.

AI agents do not have predictable patterns. They context-switch between tasks. They access multiple systems in parallel. They escalate their own capabilities through tool calls. They operate continuously. Traditional IAM systems were not built for entities that behave this way. According to Gartner (February 2026), 40 percent of enterprise applications will feature AI agents by 2026, yet only 6 percent of organizations report having an advanced AI security strategy.

That ratio, 40 percent adoption against 6 percent advanced security, is the structural vulnerability. The technology exists. Every major IAM vendor has agent identity on its roadmap. The failure is organizational: organizations are deploying AI agents under cybersecurity policies written for human users, managed by governance teams that do not own the cybersecurity function.

The Forrester prediction that an agentic AI breach will cause executive terminations in 2026 follows directly from this structural mismatch. When the breach happens, it will not be because the technology failed. It will be because the organizational structure created a seam between AI governance and cyber defense, and the attacker found it.

The Phishing Paradox

IBM’s X-Force research on AI-generated phishing reveals a nuance that most coverage of AI threats misses entirely.

An AI can generate a complete phishing campaign in five minutes using five prompts. A human social engineer needs roughly 16 hours to produce a comparable campaign. The speed advantage is real and significant.

The click-through rates tell a different story. Human-crafted phishing achieved a 14 percent click-through rate. AI-generated phishing achieved 11 percent. Humans are still better at social engineering than machines.

This is not a comforting finding. It is a terrifying one.

The economics of phishing are volume-driven. A human social engineer spending 16 hours on a campaign that converts at 14 percent produces a certain number of compromised accounts. An AI generating campaigns at 192 times that speed, even at a lower conversion rate, produces vastly more compromised accounts per unit of time. The 3-percentage-point effectiveness deficit is overwhelmed by the speed advantage.

But the paradox contains a strategic insight. AI attacks are faster but less sophisticated. They rely on scale rather than precision. This means that governance controls designed for targeted, sophisticated social engineering (training employees to spot suspicious intent) may be less effective than controls designed for high-volume, lower-quality attacks (automated URL filtering, sender verification, quarantine policies).

The cybersecurity team knows how to deploy those volume-based controls. The AI governance team understands the capability curve of the models generating these attacks. Neither team alone has the full picture. Separate them organizationally and the defense posture degrades.

Five Risks That Exist Nowhere on the Old Org Chart

McKinsey’s five agentic risk categories are worth examining individually because each one sits in the seam between AI governance and traditional cybersecurity.

Chained vulnerabilities. When agents call other agents, a vulnerability in one can propagate through the chain. This is a cybersecurity problem (vulnerability management) and an AI governance problem (agent orchestration controls) simultaneously. Neither team owns it alone.

Cross-agent task escalation. An agent designed to read files requests write access from another agent. The second agent grants it because the request appears legitimate within its context. No human approved the escalation. This is privilege escalation (cybersecurity) enabled by agentic autonomy (AI governance).

Synthetic-identity risk. AI agents can create convincing synthetic identities for authentication. Is this identity fraud (cybersecurity) or model misuse (AI governance)? It is both. Signicat’s data shows deepfake attacks grew from 0.1 percent to 6.5 percent of all fraud attempts since 2022. The absolute numbers are still small. The trajectory is not.

Untraceable data leakage. Agents process data across systems. When sensitive data appears somewhere it should not be, reconstructing the path requires understanding both the data flow (cybersecurity) and the agent’s decision logic (AI governance). One team has the logs. The other has the context.

Data corruption propagation. An agent ingests corrupted data and acts on it. Downstream agents ingest the output and propagate the corruption further. By the time anyone notices, the corruption has cascaded through multiple systems. Detection requires monitoring (cybersecurity). Containment requires understanding agent dependencies (AI governance).

Every one of these risks falls between two teams on the org chart. The space between those teams is where breaches happen.

The Framework That Already Exists

NIST’s Cyber AI Profile is not just a theoretical exercise. It maps specific AI risks to existing cybersecurity controls and identifies where new controls are needed. The framework recognizes what practitioners have been discovering the hard way: you cannot secure AI systems with cybersecurity tools alone, and you cannot govern AI systems without cybersecurity infrastructure.

ISO 27001, NIST CSF, and SOC 2 do not account for autonomous agents. McKinsey’s research is explicit on this point. The compliance frameworks that most organizations rely on were written for a world where all actors in a system are human or operate under direct human supervision. Agents operate in neither category.

This does not mean existing frameworks are useless. It means they are incomplete. The Cyber AI Profile attempts to fill that space by creating a unified control set. Organizations waiting for the final version before acting are making the same mistake organizations made with cloud security in 2014: waiting for the standard while deploying the technology.

What Convergence Looks Like in Practice

The practical question is not whether AI governance and cybersecurity should converge. They are converging regardless. The question is whether that convergence happens through organizational design or through incident response.

Convergence through design means three things.

First, unified risk assessment. Every AI deployment gets evaluated for both AI-specific risks (model behavior, training data integrity, prompt injection susceptibility) and traditional cyber risks (access control, data protection, network exposure). One assessment. One team. One report.

Second, shared control frameworks. Agent identity management uses the same infrastructure as human identity management, extended with agent-specific controls: behavioral baselines, capability boundaries, interaction logging, and automatic containment. The cybersecurity team maintains the infrastructure. The AI governance team defines the policies. They report to the same leader.

Third, integrated incident response. When an AI agent behaves anomalously, the response playbook covers both possibilities: the agent was compromised (cyber incident) or the agent is behaving as designed in an unintended way (governance incident). The initial response is identical: contain, investigate, remediate. The root cause analysis diverges, but the first 30 minutes should not depend on whether the SOC or the AI governance team picks up the alert first.

Convergence through incident response means none of this exists, and the organization discovers the need for it during a breach. CrowdStrike’s finding that 76 percent of organizations struggle to match AI attack speed suggests which version most organizations will experience.

The $1.88 Million Argument

Return to IBM’s numbers. The $1.88 million per-incident savings from AI-integrated security operations is the economic proof that convergence pays for itself. But the number has a prerequisite that most organizations overlook.

The savings do not come from buying AI security tools. They come from having the governance structures that allow AI tools to operate effectively within security operations. An AI tool running inside a well-governed security operation catches threats faster, triages more accurately, and contains incidents before they spread. The same tool running inside an ungoverned operation produces noise, false confidence, and the occasional catastrophic miss.

The average breach cost across all organizations in IBM’s study was $4.88 million. Healthcare breaches averaged $9.77 million. These are not abstract numbers. They are the cost of organizational structures that separate functions the attackers have already unified.

The Uncomfortable Vendor Question

A responsible analysis of this convergence must acknowledge the incentive structures at work.

Every data point cited in this article comes from an organization with a commercial interest in the conclusion. IBM sells security services. McKinsey sells cybersecurity consulting. CrowdStrike sells endpoint protection. Gartner sells advisory subscriptions. Anthropic sells AI models. Each of them benefits from the narrative that AI and security are converging.

That does not make their data wrong. IBM’s Cost of a Data Breach report is 19 years old and based on 604 real incidents. McKinsey’s risk taxonomy reflects genuine client engagements. CrowdStrike’s speed data comes from their own detection telemetry. These are real observations from organizations with real data.

But the vendor-fear-product cycle is real. A vendor identifies a threat, publishes research about the threat, and then sells the solution to the threat. Awareness of this cycle does not invalidate the threat. It does mean that every claim deserves scrutiny proportional to the vendor’s commercial interest in the conclusion.

The convergence of AI governance and cybersecurity is happening regardless of what any vendor says about it. The evidence is structural, not promotional. Standards bodies, compliance frameworks, and incident patterns all point in the same direction. But which specific products, services, and consulting engagements organizations need is a different question. And on that question, every source cited here has a financial interest.

What This Means Now

Organizations maintaining separate AI governance and cybersecurity functions should ask three questions.

Can either team respond to an AI-enabled attack without the other? If the answer is no, the separation is a liability.

Does the AI governance team have visibility into the security operations center? Does the cybersecurity team have visibility into AI agent deployments? If either answer is no, there is a blind spot at exactly the point where the next generation of attacks will land.

When the inevitable breach happens, will the post-mortem identify “lack of coordination between AI governance and cybersecurity” as a contributing factor? If the answer is probably yes, the time to reorganize is before the breach, not after it.

The org chart still separates these functions. The attackers do not. That asymmetry is the vulnerability, and no amount of tooling fixes an organizational design problem.


This analysis synthesizes IBM Cost of a Data Breach Report 2024 (July 2024), McKinsey’s Deploying agentic AI with safety and security (October 2025) and AI is the greatest threat and defense in cybersecurity today (May 2025), Anthropic GTG-1002 Report (November 2025), NIST Cyber AI Profile NISTIR 8596 (December 2025), and Gartner AI Agent Security Predictions (February 2026).

Victorino Group helps organizations unify AI governance and cybersecurity before the breach forces the conversation. Let’s talk.

If this resonates, let's talk

We help companies implement AI without losing control.

Schedule a Conversation