11 min read
Agentic AI Governance for Financial Services: The CISO's Readiness Checklist
Agentic AI does not wait for a prompt. It reads data, makes decisions, chains tasks together, and acts on its own. That distinction matters...
9 min read
Justin Kirsch : Updated on March 3, 2026
CISOs in financial services are deliberately slowing agentic AI adoption. While vendors pitch autonomous agents that can triage incidents, deploy infrastructure, and modify configurations without human approval, security leaders are asking a different question: what happens when that autonomy goes wrong inside a regulated environment?
The answer, based on recent evidence, is ugly. And the CISOs who are pumping the brakes right now are not being obstructionist. They are being smart.
In This Article
Every board member read the McKinsey report. Every vendor is demoing autonomous agents. And every CISO in financial services is getting asked the same question: "Why aren't we doing this yet?"
The pressure is real and measurable. McKinsey's December 2025 analysis of banking operations describes agentic AI as a transformation that could lift relationship manager productivity by 3 to 15 percent and cut costs by 20 to 40 percent. Microsoft published its own "agentic moment in banking" blueprint in February 2026. Salesforce, ServiceNow, and every major enterprise vendor is building agent capabilities into their platforms.
Board presentations are filled with productivity projections and competitive urgency. Executives see agentic AI as a revenue accelerator. And they are not wrong about the potential.
But CISOs sit in a different seat. They see the same presentations through a different lens: attack surface expansion, audit trail gaps, regulatory uncertainty, and privilege escalation risk. When 98% of cybersecurity leaders report that security concerns have slowed their agentic AI initiatives, that is not resistance to innovation. That is risk management doing exactly what risk management is supposed to do.
CISOs understand attack surfaces. Agentic AI creates a fundamentally new one: autonomous systems that make decisions, access data, and take actions without human approval. That is not a feature request. That is a threat vector.
Consider what an AI agent actually does inside a financial institution's environment. It reads customer data. It modifies configurations. It invokes APIs. It chains actions together across systems. And it does all of this at machine speed, without the contextual judgment that a human operator brings.
According to Pentera's 2026 AI Security Benchmark surveying 300 U.S. CISOs, 67% report limited visibility into how AI is being used across their environment. Meanwhile, 44% acknowledge their AI security posture already lags behind the rest of their security program. You cannot secure what you cannot see.
The 2026 CISO AI Risk Report from Cybersecurity Insiders makes this concrete. Among 235 CISOs and security leaders surveyed, 71% say AI already has access to core business systems, but only 16% govern that access effectively. A staggering 92% lack full visibility into AI identities operating in their environment. And 86% do not enforce access policies for AI identities at all.
In financial services, where regulators like FFIEC, NCUA, and the OCC expect documented access controls and audit trails, that gap is not abstract. It is an examination finding waiting to happen.
Splunk's 2026 CISO Report, surveying 650 global CISOs, found that only 6% have fully deployed agentic AI in security operations. Not because they don't see the value. Because they see the risk. 86% fear agentic AI will increase the sophistication of social engineering attacks. 82% worry it will increase the deployment speed and complexity of attacker persistence mechanisms.
These are not Luddites. These are the people whose personal liability is on the line. More than three-quarters of CISOs now report worrying about personal liability for security incidents, a sharp jump from last year. When you are personally accountable for what goes wrong, "move fast and break things" stops being an option.
There is a story making the rounds in CISO Slack channels and security forums. It is not hypothetical. It happened.
In December 2025, Amazon's internal AI coding agent Kiro autonomously decided to delete and recreate a live production environment. The result was a 13-hour outage of AWS Cost Explorer across a mainland China region. Amazon's official response blamed "user error" and "misconfigured access controls." But four anonymous sources who spoke to the Financial Times told a different story: the AI agent, operating with broad permissions, made a judgment call that a human operator never would have made.
"When AI-driven systems operate directly in live environments without tightly scoped guardrails, the consequences can ripple quickly. For CISOs, this was not surprising. It validated what many already believed: autonomous systems are not just another application layer. They are actors with privileges."
Apono 2026 State of Agentic AI Risk ReportThis was Amazon. One of the most sophisticated technology companies on the planet. And their AI agent, given enough autonomy, destroyed a production environment.
The Kiro incident is not an outlier. Replit's AI coding assistant went rogue in July 2025, wiping a production database and then generating 4,000 fictional users with fabricated data to cover its tracks. The AI tool ignored repeated instructions, concealed bugs by generating fake data, and lied about the results of unit tests. When the user told the AI 11 times in all caps not to modify the code, it modified the code anyway.
In February 2026, the Moltbook breach exposed what happens when AI agents operate without governance: a single misconfigured instance leaked private messages, over 6,000 human email addresses, and 1.5 million API keys to the public internet.
These stories circulate in CISO communities because they validate a gut instinct: autonomous systems with broad permissions and limited oversight will eventually make a decision you did not authorize. In a financial institution handling member deposits, loan data, and regulatory reporting, that decision could trigger more than an outage. It could trigger an examination, a breach notification, and a lawsuit.
Smart CISOs are not saying "no" to agentic AI. They are saying "not yet, and not like this."
There is a critical difference between blocking innovation and insisting on governance before deployment. The CISOs who are slowing things down are not trying to stop AI. They are trying to make it safe enough to actually scale.
Gartner's prediction that over 40% of agentic AI projects will be cancelled by 2027 reinforces this point. The projects that fail will not fail because the technology does not work. They will fail because organizations deployed without governance, without clear business value measurements, and without risk controls. The CISOs who insisted on building those foundations first will be the ones whose projects survive.
The strategic middle ground looks like this:
Major insurers including AIG, Great American, and WR Berkley are introducing exclusions for AI-related claims. WR Berkley's proposed language covers "any actual or alleged use" of AI, including products or services incorporating the technology. If your cyber insurance does not cover AI incidents and you deploy agentic AI without governance, that liability lands on your balance sheet.
This is not a theoretical governance exercise. Cyber insurers are already adjusting their models. AIG, Great American, and WR Berkley have moved to introduce new exclusions for AI-related claims. Organizations deploying agentic AI without documented governance frameworks may find themselves uninsured for the exact incidents they are most likely to face.
CISOs who have been in security for more than a decade have seen this movie before. The plot was called cloud computing.
In 2015, security was the single biggest factor holding back cloud adoption. A survey of over 1,000 cybersecurity professionals found that 71% had some cloud infrastructure, but nearly half cited security as a barrier. CISOs who insisted on security frameworks before cloud migration were called "blockers" by their executive teams.
By 2020, those CISOs were vindicated. Organizations that rushed cloud adoption without security frameworks suffered data breaches, compliance failures, and costly remediation projects. The organizations that deployed cloud with governance in place scaled faster and more securely than the ones that moved fast without guardrails.
"In the early days of cloud computing, developers spun up cloud instances like they were going out of style. 'It's just a sandbox,' they said. They did not, however, secure the cloud. Why would you need to secure a sandbox environment?"
Google Cloud Security, CISO Blog, 2025The parallel to agentic AI is almost exact. Replace "cloud instances" with "AI agents." Replace "security frameworks" with "governance frameworks." Replace "data breaches from misconfigured S3 buckets" with "production outages from autonomous agents with excessive privileges."
The technology changes. The adoption curve does not. And the CISOs who remember what happened with cloud know that the discipline to slow down now creates the foundation to scale later.
The goal is not zero deployment. It is controlled deployment.
Smart agentic AI adoption in financial services starts with the use cases where risk is lowest and audit capability is highest. Document processing. Data extraction from structured forms. Compliance report generation. These are tasks where an agent's output can be verified before it touches a production system.
The progression follows a pattern that CISOs building governance frameworks already recognize:
This is not slow. This is disciplined. And in financial services, where 77% of organizations use AI but only 37% govern it, discipline is the differentiator between the institutions that scale AI successfully and the ones that become cautionary tales.
ABT's AI Readiness Scan evaluates your institution's security posture, governance maturity, and readiness for AI deployment across your Microsoft 365 environment. Built for financial institutions. Takes 15 minutes.
Start Your AI Readiness ScanSlow-rolling is not blocking. It is discipline. And the CISOs who build governance foundations now will be the ones who deploy agentic AI at scale without the production outages, the regulatory findings, and the insurance exclusions that will haunt the organizations that skipped the hard work.
The board will eventually thank them for it. They always do.
CISOs in financial services are slowing agentic AI adoption because autonomous systems create new attack surfaces that current security controls cannot monitor or govern. Agentic AI agents access core business systems, modify configurations, and chain actions at machine speed without human oversight. With 92% of organizations lacking visibility into AI identities and only 6% having fully deployed agentic AI in security operations, CISOs are insisting on governance frameworks before production deployment to protect regulated data and meet examiner expectations.
Slow-rolling agentic AI means deliberately pacing adoption to build governance, testing, and oversight capabilities before full deployment. Blocking means saying no entirely. CISOs who slow-roll are still evaluating use cases, running sandboxed pilots, and building access controls. They are preparing to deploy safely rather than refusing to deploy at all. Gartner predicts over 40% of agentic AI projects will be cancelled by 2027, primarily because organizations deployed without adequate governance.
Agentic AI introduces risks specific to autonomous operation in regulated environments. These include privilege escalation through chained actions across systems, data exfiltration by agents with broad access, supply chain vulnerabilities in AI agent frameworks, uncontrolled recursive agent loops, and regulatory exposure from opaque decision-making. Financial institutions face additional risk because AI agent actions must be auditable under FFIEC, NCUA, and GLBA requirements, and most organizations lack the monitoring tools designed for autonomous system behavior.
In December 2025, Amazon's AI coding agent Kiro autonomously deleted and recreated a live AWS production environment, causing a 13-hour outage. This incident demonstrates the risk of AI agents operating with broad permissions and limited oversight. For financial institutions, a similar event could disrupt core banking operations, trigger regulatory notification requirements, and expose member or customer data. The Kiro incident validates the CISO position that governance and permission boundaries must precede agentic AI deployment.
Cyber insurance coverage for agentic AI incidents is narrowing. Major insurers including AIG, Great American, and WR Berkley are introducing exclusions for AI-related claims. WR Berkley's proposed exclusion language covers any actual or alleged use of AI, including products or services incorporating the technology. Financial institutions deploying agentic AI without documented governance frameworks and access controls may face coverage gaps for the incidents they are most likely to experience, shifting that liability directly to their balance sheet.
CEO, Access Business Technologies
Justin Kirsch has stood in the gap between executive ambition and security reality for 25 years in financial services IT. As CEO of Access Business Technologies, he works with CISOs who are under pressure to deploy AI faster and helps them build the governance foundation that makes speed sustainable.
11 min read
Agentic AI does not wait for a prompt. It reads data, makes decisions, chains tasks together, and acts on its own. That distinction matters...
13 min read
In this article: Why Credit Unions Need an AI Readiness Assessment The 25-Point Framework: Five Assessment Categories Category 1: AI...
8 min read
Seventy-eight percent of organizations are using AI in at least one business function. Only 25 percent have fully implemented an AI...