Your Company’s New ‘AI Agent’ Isn’t an Efficiency Tool—It’s a Shadow CFO With a Spending Limit (and auditors are already freaking out)
Once software can initiate transactions, the risk stops being “bad advice” and becomes unauthorized execution. Auditors don’t want vibes—they want authority, logs, ownership, and testable controls.

Key Points
- 1Treat agentic AI like delegated authority: once it initiates transactions, the risk becomes control bypass, not just bad outputs.
- 2Reframe permissions around money: spending limits can help, but only with immutable logs, ownership, monitoring, and testable evidence.
- 3Expect audit and regulatory pressure: COSO’s 2026 GenAI controls and FINRA’s deepfake warnings raise the bar for approvals and records.
A finance leader used to worry about spreadsheets.
Now the anxiety sits somewhere else: inside the workflow.
An “AI agent” that executes starts looking like delegated finance
That’s why a quiet phrase has begun to circulate in audit and internal-control circles: shadow CFO. Not because a bot is plotting in the dark, but because autonomy changes the risk. The failure mode shifts from “bad advice” to unauthorized execution, and the consequences arrive faster than most control environments were built to handle.
“The moment software can initiate transactions, the conversation stops being about accuracy and starts being about authority.”
— — TheMurrow
The new meaning of “agentic”: autonomy that touches money
Deloitte’s internal audit “hot topics” report for 2026 goes further, flagging a specific category of exposure: “control bypass risk.” In plain English, an agent can initiate actions outside established approval chains—raising the prospect of fraud, financial misstatements, or data breaches. That is not a philosophical objection to AI. It is a map of how modern systems fail when speed outruns oversight.
From “tool” to “operator” in one integration
Agentic systems introduce a different dynamic. When a system can decide which step comes next—what invoice to route, what exception to clear, which vendor to select—it becomes hard to describe the boundary between “automation” and “delegation.” Engineers often focus on whether the agent completes the task. Auditors focus on whether the agent completed the task with the right authority, evidence, and accountability.
Why the “spending limit” is becoming the new permission boundary
Call it a workaround or a mature control design—either way, it reflects a new reality: spend thresholds are becoming a proxy for human approval. That can be sensible. It can also create blind spots if limits are set loosely, monitored weakly, or treated as a substitute for attribution and audit evidence.
“Spending limits look like safety rails—until you realize the rails are meaningless without logs, ownership, and testing.”
— — TheMurrow
What auditors are actually worried about: evidence, not vibes
A landmark in that translation arrived with COSO’s “Achieving Effective Internal Control Over Generative AI (GenAI)” (©2026). COSO is not a startup with a point of view; it’s the organization behind the Internal Control–Integrated Framework that underpins much of corporate control design and external assurance. The message is direct: rapid GenAI adoption can compress decision cycles and introduce risks that threaten reliability of operations, reporting, and compliance if unmanaged.
COSO’s shift: from principles to audit-ready templates
- A capability-first taxonomy of GenAI uses (eight types, including “orchestration” and “judgment”) with tailored control implications
- Minimum control expectations aligned to COSO’s five components
- Templates for risk assessment matrices, control testing procedures, and metrics dashboards—materials designed to close the gap between broad governance talk and what auditors need as evidence
That last point is the hinge. When auditors request proof, they are not asking for a slide deck about “responsible AI.” They are asking for documentation they can test: who approved the design, what access controls exist, what logs are retained, what monitoring runs, and what changed when.
Four risk classes COSO explicitly calls out
1. Prompt-injection attacks (inputs that manipulate a model into taking unsafe actions)
2. Hallucinations (plausible but incorrect outputs)
3. Opaque reasoning (difficulty explaining why a decision was made)
4. Model drift and rapid configuration changes (behavior changes over time)
None of these are abstract in finance. An “orchestration” use case that can move from recommendation to execution turns each category into potential books-and-records pain.
“Auditors aren’t panicking about AI. They’re reacting to how quickly small control gaps can become material when execution is automated.”
— — TheMurrow
“Who’s accountable?” The attribution problem when an agent acts
Accountability in finance is not an ethical preference. It is a control requirement. External auditors, regulators, and boards expect a company to answer questions that sound simple until an agent sits in the middle:
- Who approved the transaction?
- Who had access to initiate or alter it?
- Who changed the system configuration?
- What evidence supports reliance on automated controls?
Delegation without ownership becomes “shadow finance”
Combine shadow AI with agentic execution and you get the organizational version of the “shadow CFO” idea: spend decisions and workflow approvals start happening in parallel to formal finance operations. Not necessarily malicious. Often it begins as a well-meaning team automating exceptions to move faster. The problem arrives when those exceptions become routine and nobody can say, with confidence, what the agent is authorized to do today.
Audit evidence is a design problem, not a documentation problem
- Durable logs showing prompts, actions, and outcomes
- Controls over who can change the agent’s tools, permissions, or thresholds
- Monitoring that detects unusual vendor changes or payment patterns
- A clear line of responsibility for ongoing performance and incident response
Without those, accountability becomes a debate. Auditors dislike debates.
The regulatory thread: deepfakes weaken approvals as automation expands
FINRA’s 2026 Annual Regulatory Oversight Report (dated Dec. 9, 2025) adds a standalone GenAI section (“NEW FOR 2026”) and reiterates risks tied to cyber-enabled fraud and identity deception—voice clones, fake IDs, deepfake selfies. That matters for finance operations because the choke points in procure-to-pay and treasury often rely on human verification: call-backs, email approvals, identity checks, and exception handling.
A dangerous overlap: weaker authentication, faster execution
1. Impersonation is improving. Deepfakes and voice cloning raise the odds that an approval request looks and sounds legitimate.
2. Workflows are accelerating. Agentic automation reduces friction between request and execution.
Agents do not need a literal “corporate card” to create spend risk. If the approval environment becomes easier to spoof, then pushing more execution into automated systems increases exposure—especially if the agent can act on messages, tickets, calls, or loosely validated identities.
“Books and records” pressure arrives through the side door
If the organization can’t reconstruct an agent’s actions—or can’t show why it trusted the agent to take them—then the incident becomes more than a security failure. It becomes a governance failure. COSO’s emphasis on audit-ready mapping anticipates that reality.
Where the money actually moves: procure-to-pay and treasury as the front line
Procure-to-pay: speed creates new failure modes
- Creating or modifying purchase orders
- Selecting vendors or switching vendors during sourcing
- Approving invoices or clearing exceptions
- Routing approvals based on inferred policy
- Triggering payment actions once conditions appear “met”
Each touchpoint has a classic control objective—authorization, segregation of duties, completeness, accuracy. Agentic orchestration compresses these steps into fewer moments of human review, which raises the stakes for the controls that remain.
Treasury: reversibility matters more than cleverness
That’s why finance leaders increasingly ask a blunt operational question: Is the action reversible? If the answer is “no” or “not quickly,” then autonomy needs tighter boundaries, stronger authentication, and better monitoring than many early agent deployments include.
Practical control design: what “good” looks like without killing innovation
COSO’s 2026 guidance is helpful here because it translates principles into control expectations. Deloitte’s warnings are useful because they identify likely failure patterns: control bypass, shadow deployments, and unclear accountability. Put together, they point to a workable agenda.
Minimum viable controls for agentic finance workflows
- Defined authorization boundaries (what the agent can do, up to what limits, in which systems)
- Spend thresholds and policy checks enforced technically, not just documented
- Segregation of duties preserved through permissions (an agent should not both create a vendor and approve a payment to it)
- Immutable logging of prompts, actions, approvals, and system changes
- Change control for agent configuration, tools, and access rights
- Monitoring and exception alerts (unusual vendor changes, payment spikes, out-of-policy actions)
- Human review for high-risk actions (irreversible payments, vendor bank-detail changes, budget transfers)
These are not exotic. They are the internal-control playbook translated into agentic terms.
The cultural change: finance needs a seat in the build, not just the audit
The answer is not to make finance the department of “no.” The answer is to embed control requirements early: finance and internal audit define the control objectives, security defines identity and access, IT defines logging and change control, and the business defines the workflow. Agents then operate inside a governed system rather than alongside it.
A fair counterpoint: autonomy can strengthen controls—if designed honestly
Humans bypass controls too. Humans approve invoices while distracted. Humans can be bribed, coerced, or socially engineered. A well-designed agent can enforce policies consistently, apply thresholds reliably, and produce richer audit trails than a scattered email-based approval process.
Deloitte’s warnings do not argue against autonomy. They argue against autonomy that bypasses controls. COSO’s guidance does not argue against GenAI. It argues for internal control design that keeps pace with compressed decision cycles.
The “shadow CFO” framing lands because it captures a true shift: when systems can execute, they hold financial power. Whether that power reduces risk or multiplies it depends on whether the organization treats agentic capability as a finance transformation—or as a clever plugin.
A mature view accepts both realities: agentic AI can improve compliance through consistent enforcement, and it can also create new single points of failure if accountability, access, and evidence are not engineered from the start.
TheMurrow takeaway: the fight isn’t about AI—it’s about authority
The question is whether the organization can answer, with evidence, the classic control questions in an agentic world: Who authorized it? Who can change it? What happened? Can you prove it?
A “shadow CFO” is not a robot in a corner office. It’s a system acting with financial consequence while governance lags behind. COSO’s 2026 work makes the control expectations explicit. Deloitte’s 2026 audit warnings make the bypass risks explicit. FINRA’s 2026 focus on GenAI-enabled deception makes the approval environment weaker at the same time automation is getting stronger.
Finance leaders don’t need to fear autonomy. They need to price it correctly: autonomy without controls is not efficiency. It’s an unbudgeted liability.
1) What is an “AI agent” in finance, exactly?
2) Why are auditors more concerned about agents than chatbots?
3) What does “control bypass risk” mean?
4) What is “shadow AI,” and why does it matter?
5) How does GenAI-enabled fraud affect finance approvals?
6) Are spending limits a real solution—or just a bandage?
7) Can agentic AI improve internal controls rather than weaken them?
Key Insight
Minimum viable controls (as described in the article)
- ✓Defined authorization boundaries (what the agent can do, up to what limits, in which systems)
- ✓Spend thresholds and policy checks enforced technically, not just documented
- ✓Segregation of duties preserved through permissions
- ✓Immutable logging of prompts, actions, approvals, and system changes
- ✓Change control for agent configuration, tools, and access rights
- ✓Monitoring and exception alerts (unusual vendor changes, payment spikes, out-of-policy actions)
- ✓Human review for high-risk actions (irreversible payments, vendor bank-detail changes, budget transfers)
Frequently Asked Questions
What is an “AI agent” in finance, exactly?
Deloitte describes agentic AI as systems that can autonomously plan, execute, and adapt actions to achieve objectives, often by connecting across tools and initiating tasks. In finance, that can mean moving beyond drafting analysis to taking operational steps—routing approvals, creating purchase orders, or triggering payments—depending on permissions and integrations.
Why are auditors more concerned about agents than chatbots?
Chatbots mostly create advice and content. Agents can execute actions. That shifts the risk from “wrong output” to unauthorized or unreviewed transactions, which directly affects internal controls over operations, financial reporting, and compliance. COSO’s 2026 GenAI guidance reflects this by mapping risks and controls in audit-ready terms.
What does “control bypass risk” mean?
Deloitte uses the term to describe situations where an AI agent initiates actions outside established approval chains. A control bypass can happen through excessive permissions, unclear boundaries, shadow deployments, or weak monitoring. The risk is not theoretical: bypassed approvals can lead to fraud exposure, misstatements, or data breaches.
What is “shadow AI,” and why does it matter?
“Shadow AI,” flagged by both Deloitte and COSO, refers to GenAI systems deployed by business units without enterprise oversight. In finance workflows, that can create a parallel, informal decision-and-execution channel—hard to monitor, hard to audit, and difficult to hold accountable when incidents occur.
How does GenAI-enabled fraud affect finance approvals?
FINRA’s 2026 Annual Regulatory Oversight Report highlights GenAI-enabled deception such as voice clones, fake IDs, and deepfake selfies. Those techniques can weaken the reliability of identity checks and approval workflows. If approvals become easier to spoof while agentic systems make execution faster, the organization’s exposure increases.
Are spending limits a real solution—or just a bandage?
Spending limits can be a practical control boundary—caps per transaction, per day, or per vendor—especially when agents are allowed to execute routine tasks. Limits are not sufficient by themselves. Auditors also expect logs, change control, access governance, monitoring, and clear accountability for agent behavior and configuration.















