Why 62% of Organizations Still Cannot Scale AI Safely
Key Highlights
- Most organizations cite security and risk management as the primary barrier to scaling AI, not technical or regulatory issues.
- Effective AI governance requires treating AI as an infrastructure project with dedicated identity, access control, and audit layers from the outset.
- CISOs should involve security leaders early in AI strategy discussions to embed controls before deployment, reducing operational debt.
- Demonstrating compliance through tamper-evident audit trails and policy enforcement reports is key to gaining board confidence.
- Operationally, AI agents must be assigned distinct credentials, with continuous monitoring and automated deprovisioning to contain incidents quickly.
I am the CISO. That means I sit between a CEO who wants AI agents deployed across every business unit by year-end and a board that wants to know those agents will not become the next breach headline. Those two demands land on the same desk, mine, and for the past year, reconciling them has been the hardest part of my job.
The Stanford 2026 AI Index Report just confirmed that I am not alone. Sixty-two percent of organizations cite security and risk as the number-one barrier to scaling agentic AI. Not technical limitations. Not regulatory uncertainty. Not budget. Security — by a 24-point margin over every other factor.
That number tells me two things. First, most CISOs are in the same position I am. Second, the ones who figure out how to unblock AI scaling without accepting unmanageable risk will be the ones whose organizations pull ahead. The rest will spend 2026 explaining to their CEOs why the AI roadmap is stalled.
The CEO Conversation
My CEO does not think in terms of attack surfaces, control frameworks, or vulnerability classes. He thinks in terms of competitive advantages, time-to-market and revenue. When he asks me about AI, the question is never "is it secure?" The question is "when can we deploy it?"
The honest answer, “we cannot deploy it safely yet because we lack the controls to govern what these agents do with our data," is not a conversation stopper if I frame it correctly. But it does require me to translate the security barrier into business language.
Here is how I frame it now, using the Stanford data. Organizations that have already deployed AI agents without adequate controls are the ones generating the incidents Stanford documented. The share of organizations experiencing 3–5 AI incidents annually rose from 30% to 50% between 2024 and 2025. Those are not random failures; those are the aggressive adopters whose deployments outran their governance. Meanwhile, organizations rating their own AI incident response as "excellent" dropped from 28% to 18%. The organizations that move fastest are also the ones handling incidents the worst.
Organizations that have already deployed AI agents without adequate controls are the ones generating the incidents Stanford documented. The share of organizations experiencing 3–5 AI incidents annually rose from 30% to 50% between 2024 and 2025.
That is the cautionary story my CEO needs to hear. Not "AI is risky,” he knows that. The story is: "organizations that deployed AI without the security architecture to govern it are now spending more time on incident response than on AI innovation. We can avoid that if we build the controls first."
The Board Conversation
The board conversation is different. The board does not want to hear about my IR playbooks. The board wants to know three things: our exposure, what we are doing about it, and whether we can demonstrate it to regulators and auditors.
Stanford's data gives me a framework for answering all three.
On exposure: 62% of organizations cannot scale agentic AI because of security concerns. AI agent unguided solve rates on cybersecurity tasks rose from 15% to 93% in twelve months on the Cybench benchmark — meaning adversaries can now automate the same operations my red team runs manually. The Agents of Chaos study, conducted at Harvard, MIT, Stanford, and Carnegie Mellon, demonstrated that AI agents can be fully compromised through conversation alone — no exploit code required.
On what we are doing about it: I describe the four operational changes I have made. I am rewriting IR playbooks for AI-specific scenarios. I am treating every AI agent as a privileged identity with its own credentials, access scope, and audit trail. I am building a kill-switch capability that works in production, not just in test environments. And I am updating my threat model to reflect adversary AI capabilities at the Cybench 93% level, not the human-speed assumptions my SOC was designed around.
On demonstrability: this is where most CISOs lose the board. The board does not want to hear "we have controls." The board wants to see evidence. Tamper-evident audit trails. Policy enforcement logs. Agent access reports. If I cannot produce an evidence package showing which agents accessed which data under which policy during the last quarter, I do not have a defensible posture; I have a stated intention. Stanford found that organizations with formal responsible AI policies experienced an eight percentage-point decrease in incidents. But the gap between "we have a policy" and "we can produce evidence the policy was enforced" is the gap where board confidence lives.
The Operations Reality
The CEO and board conversations happen quarterly. The operations reality happens every day.
When an AI agent misbehaves in my environment, I do not get to write a memo about it — I have to contain it, assess the blast radius, and determine whether regulated data was affected. All of that depends on instrumentation that most organizations have not built for AI.
The most common gap I see, in my own environment and in peer conversations, is that AI agents run under shared service accounts with broad permissions. No distinct identity. No scoped authorization. No independent audit trail. When something goes wrong, the investigation starts with "which agent was it?" and the answer is "we cannot tell." That is the configuration Stanford's incident data is measuring.
When an AI agent misbehaves in my environment, I do not get to write a memo about it — I have to contain it, assess the blast radius, and determine whether regulated data was affected. All of that depends on instrumentation that most organizations have not built for AI.
The fix is the same privileged access management discipline we apply to human administrators: distinct credentials, just-in-time provisioning, session-level audit trails, and automated deprovisioning. The difficult part is extending that discipline to agents operating at machine speed across dozens of repositories simultaneously without creating so much friction that the business abandons governed AI for shadow deployments.
The 38% That Figured It Out
Stanford's 62% means 38% have cleared the barrier. Based on my own experience and conversations with peers, three patterns stand out.
First, they treated AI governance as an infrastructure project, not a policy project; building the identity layer, the access control layer, and the audit layer before deploying agents at scale. Second, the CISO had a seat at the AI strategy table from the beginning, not after the first incident. Third, they accepted that AI governance requires continuous investment. The threat model changes quarterly. Agent capabilities change quarterly. Controls must keep pace.
The organizations stuck in the 62% are the ones that treated AI as an IT project and security as a tollgate at the end. By the time the CISO is involved, the agents are in production, the shared service accounts are configured, and the operational debt is too deep to fix without disrupting the business.
What I Would Tell Any Peer CISO This Week
If the Stanford data lands on your desk and your CEO asks what it means, do not lead with the risk. Lead with the opportunity. 62% of our competitors cannot scale AI because they lack a security architecture. Here is our plan to be in the 38% that can."
Then deliver the plan. Rewrite the IR playbook for AI-specific scenarios. Treat every agent as a privileged identity. Build a kill-switch capability and test it under realistic conditions. Update the threat model for adversary AI at the Cybench 93% level. And produce evidence, not assurances, that your controls are enforced.
Stanford just measured the gap. Closing it is our job.
About the Author
Frank Balonis
Chief Information Security Officer and Senior VP of Operations and Support at Kiteworks
Frank Balonis is Chief Information Security Officer and Senior VP of Operations and Support at Kiteworks, with more than 20 years of experience in IT support and services. Since joining Kiteworks in 2003, Frank has overseen technical support, customer success, corporate IT, security, and compliance, collaborating with product and engineering teams. He holds a Certified Information Systems Security Professional (CISSP) certification and served in the U.S. Navy. He can be reached at [email protected].


