US Banks Deploy AI Agents to Counter AI-Powered Fraud

US Banks Deploy AI Agents to Counter AI-Powered Fraud

Behind the smooth tap-to-pay and instant transfers that consumers take for granted, an arms race has intensified as AI-empowered fraud rings use autonomous tools to probe bank defenses, learn from outcomes, and pivot faster than legacy controls can respond, forcing U.S. financial institutions to deploy their own AI agents that monitor, decide, and act at machine speed to keep payments safe. At the Smarter Faster Payments conference in San Diego, technologists, anti-fraud strategists, and cloud providers described a decisive shift: rules-based systems and slow model refreshes are yielding to agentic AI, continuous monitoring, and tighter collaboration with cloud platforms. The conversation extended beyond tooling to operating models and supervision, signaling that speed, auditability, and explainability must be engineered together, not traded off, if banks are to keep pace with industrialized attacks.

The New Threat Landscape

Fraud no longer resembled a string of isolated scams; it operated like a software business with pipelines, testing cycles, and automated distribution across channels. Speakers warned that “polymorphic agentic agents” reconfigured behavior mid-attack when encountering new controls, cycling through identity artifacts, device fingerprints, and conversational patterns until something worked. Matt Vega of Sardine described onboarding barrages that quickly abandoned a blocked vector and reappeared via mule networks or altered payloads. Call centers faced AI-generated voice floods that spoofed caller metadata while social engineering scripts adapted on the fly. In this environment, static thresholds and quarterly model updates became liabilities rather than safeguards.

The speed gap showed up most painfully in payments operations, where attackers studied rail-specific timing, settlement windows, and exception handling. They chained small advantages—like exploiting differences between batch and real-time posting—into high-velocity raids. Greg Williamson of Nasdaq Verafin cautioned that criminal testing loops now run in hours, not months, meaning a model’s theoretical lift mattered less than how quickly it could be deployed, observed, and recalibrated. Vendors added that adversaries increasingly targeted platform architecture itself, looking for brittle integrations or unmonitored microservices. Once inside a workflow, agentic systems varied payloads to avoid signature-based filters, then fanned out across lookalike channels designed to diffuse detection.

Why AI Versus AI Is Inevitable

Panelists framed the moment bluntly: human-led rules could not match an opponent that tested, learned, and relaunched in near real time. “Model quality without agility is theater,” Williamson said, noting that criminals measured bank responses and tuned their agents accordingly. Vega added that attackers’ agentic systems monitored controls during live operations, switching tactics when friction or declines rose, a capability that demanded adaptive countermeasures. The consensus was that AI had moved from enhancement to necessity, with banks treating autonomous agents as front-line responders that triaged anomalies, generated hypotheses, and recommended immediate control changes under tight governance.

This inevitability also reshaped team roles. Rather than analysts combing through queues to craft new rules, institutions pushed toward orchestrators who supervised AI agents, validated reasoning, and approved targeted experiments. Feedback loops tightened. Featurespace’s Dave Excell described models that predicted customer behavior patterns to flag subtle deviations without flooding operations with false positives. That shift aimed to preserve user experience while raising attack costs. The argument was not that humans were sidelined, but that their highest-value work occurred in oversight, investigation, and policy tuning, supported by systems that could test and learn safely in production-like sandboxes.

Building Agent-Based Defenses

Banks and providers were assembling layered defenses built around autonomous agents, streaming analytics, and investigator assist tools. Elizabeth Bourgoin of Google Cloud outlined agents that continuously scanned client environments for misconfigurations, privilege creep, and control blind spots, then prioritized patching steps with evidence and risk impact. Detection bled into remediation as agents proposed control changes—a new velocity check, a modified authentication challenge, a refined device score—and routed them through model governance. Simulations ran before and after deployment to quantify lift and side effects, helping risk teams decide whether to proceed, roll back, or target a narrow customer segment.

Operationally, institutions stitched these agents into customer touchpoints, payment flows, and back-office casework. Real-time behavioral models tracked habits across devices and channels, improving anomaly detection during account opening and high-risk transfers. Investigator copilots summarized cross-channel evidence, extracted entities from unstructured notes, and suggested next steps grounded in policy. Excell pointed to live cases where AI filtered AI-generated call floods by voiceprint clustering and intent classification, protecting staffing levels. The unifying mechanics were continuous monitoring, short release cycles, and post-deployment learning, so each control change fed the next hypothesis rather than awaiting a quarterly review.

Regulatory Momentum for Responsible Agentic AI

Regulatory alignment accelerated, according to Vega, who said members of the House Financial Services Committee, alongside the OCC and FDIC, were working on a framework that expressly permitted agentic AI with guardrails. Institutions already used AI in pockets, but lacked codified clarity for broad, rapid adoption. The emerging approach emphasized governance, explainability, testing rigor, and monitoring, with bipartisan support reflecting the need to counter systemic fraud risks without stifling innovation. Executives welcomed this trajectory, arguing that clear rules reduced legal uncertainty and supported faster iteration by making supervisory expectations explicit upfront.

This momentum reached into the mechanics of deployment. Banks sought templates for pre-production testing, bias assessment, change management, and model documentation that exam teams could review consistently. Privacy and data residency rules shaped where models trained and how features were constructed. To keep pace with adversaries while preserving consumer protections, institutions aligned risk, compliance, and engineering on shared controls: human-in-the-loop thresholds for autonomous actions, red-teaming of agent behavior, and audit trails that captured decisions in plain language. The target state balanced autonomy with accountability, so rapid defenses remained traceable, reversible, and fair under scrutiny.

The Path Forward for Banks and Payment Providers

With threat sophistication rising and regulatory clarity taking shape, leaders focused on execution: compressing model approval timelines from months to days, adopting architectural patterns that separated policy from code, and investing in telemetry that let analysts observe agent reasoning instead of opaque outputs. Cloud platforms became force multipliers. Bourgoin said managed agents integrated across data lakes, messaging backbones, and identity services, accelerating detection and hardening infrastructure. Joe Fuqua of Truist noted that scale in data and talent mattered only if governance moved at the same pace, urging banks to streamline signoffs and enable reversible, low-risk launches that could be tuned in hours.

The practical blueprint emphasized three moves. First, stand up agentic detection at key choke points—onboarding, device binding, and outbound transfers—so defenses saw attacks early and constrained blast radius. Second, embed post-deployment learning: capture false positives and near misses, then auto-generate candidate features and thresholds for review. Third, treat cloud partners as strategic controls, not just hosting: standardize interfaces for alerts, patch orchestration, and simulation so tooling scaled across business lines. Together, these steps aimed to shrink the gap between attacker iteration and defender response while preserving customer experience and meeting supervisory expectations.

What Comes Next for Responsible AI Defense

The next stage demanded disciplined acceleration. Institutions needed to formalize agent playbooks that specified when autonomous actions were allowed, how overrides worked, and which metrics triggered rollback. Model risk teams could pre-approve families of changes—like velocity or authentication thresholds—so agents operated within safe bounds. Banks also benefited from investing in red-team exercises that used synthetic adversaries to stress controls, coupled with tabletop drills that trained executives on rapid decision-making when agents flagged emergent risks. Data contracts across systems reduced drift, preserving model fidelity as products evolved.

The industry trajectory, as articulated at the conference, pointed to an ecosystem response: banks deploying adaptive agents, vendors refining behavior-based analytics, and regulators codifying responsible use. To make progress concrete, institutions should appoint single-threaded owners for fraud AI, budget for continuous testing environments, and negotiate cloud service terms that include measurable response times for patching critical misconfigurations. Building shared threat libraries—with sanitized patterns of agentic attacks—would raise collective readiness. Done this way, the sector met autonomy with autonomy, and it did so with controls that were explainable, reversible, and resilient under pressure.

Subscribe to our weekly news digest.

Join now and become a part of our fast-growing community.

Invalid Email Address
Thanks for Subscribing!
We'll be sending you our best soon!
Something went wrong, please try again later