The digital transformation has led to significant advancements in authentication and identity verification technologies and other cyber defenses.
From biometrics to multi-factor authentication (MFA) to use of Artificial Intelligence (AI) enhanced detection and response tools, these systems are the first line of critical defense against unauthorized access in critical sectors such as finance, healthcare, manufacturing and government. However, with the rapid development of Multi-Modal AI and agentic AI, a new challenge has emerged—one that may compromise the very systems designed to protect us. By integrating multiple forms of data (e.g., voice, video, text) in multi-modal AI and use of agentic AI (automated decision-making with little or no human intervention), malicious actors are increasingly capable of bypassing authentication and identity verification security and other defenses, thereby posing a new level of cybersecurity threat. The rapid deployment of AI integrated into a wide variety of commercial products, platforms and workflows has dramatically expanded the potential attack surface.
Indeed, on November 13, 2025, Anthropic reported how its AI-powered Claude Code tool was leveraged for a fully automated sophisticated attack targeting large technology companies, financial institutions, manufacturing and government agencies: “We believe this the first documented case of a large-scale cyberattack executed without substantial human intervention.” Similarly, researchers recently reported the discovery of a strain of ransomware that used large language models to autonomously implement ransomware attacks by generating malicious code in real time. We have previously highlighted in our blogs the escalating threats to employees from DeepFake technologies and AI augmented phishing attacks.
What is Multi-Modal AI?
Multi-modal AI refers to systems that can process and combine information from diverse sources to understand and respond to inputs in ways that are more holistic and human-like. For example, rather than relying on just one modality, such as voice recognition or facial recognition, multi-modal systems can integrate text, video, and other sensory data for improved accuracy and flexibility. While these advancements offer immense potential in fields like healthcare and customer service, they also raise serious concerns when leveraged maliciously.
As more organizations implement biometric authentication, such as facial recognition and voice biometrics, multi-modal AI offers attackers a new arsenal for bypassing these security measures. By synthesizing data from multiple sources—such as voice recordings, photos, and even social media interactions—an attacker can create a comprehensive digital identity profile that closely mirrors the real thing. This new breed of attack can go beyond traditional hacking methods, using AI to trick systems that once seemed impenetrable.
What is Agentic AI?
Agentic AI generally refers to artificial intelligence systems that are capable of operating and developing autonomously and independently with little or no human oversight. Agentic AI may be integrated into systems through Application Programming Interfaces (APIs). Gartner reports that “[b]y 2028, 33% of enterprise software applications will include agentic AI, up from less than 1% in 2024, enabling 15% of day-to-day work decisions to be made autonomously.”
The AI-Powered Deepfake Threat
One immediate concern is the rise of AI-driven deepfakes. Deepfakes—hyper-realistic media created through AI that can mimic someone's appearance, voice, and behavior—have already made waves in the world of media and politics. However, these technologies are increasingly being adapted for malicious purposes, particularly in the realm of identity fraud.
An attacker could use multi-modal AI to create a convincing deepfake that mimics not just one, but several facets of an individual’s identity. For instance, by combining a victim’s facial data and voice samples with text-based information (like emails or social media posts), an AI could generate an extremely accurate imitation of the individual. This synthetic identity could then be used to bypass security systems, such as voice-activated banking systems, facial recognition used for mobile authentication, or even online verification processes employed by financial institutions.
As noted by the Center for Cybersecurity Policy and Law, deepfakes and other AI-powered impersonation techniques are particularly dangerous in financial services. Systems that rely on voice recognition or facial biometrics are becoming increasingly vulnerable to attacks that could potentially manipulate the very data they rely on for authentication. As acknowledged by the U.S. Treasury, AI has the capability to mimic biometrics (such as photos/video or a customer or the customer’s voice). As discussed further below, this capability is a growing concern, especially in the context of digital identities in the financial sector, where the consequences of breaches could be severe.
Erosion of Trust in Biometric Authentication
Biometric authentication, once hailed as a more secure alternative to traditional passwords, is being challenged by the rapid advancements in AI. Unlike passwords, which can be changed if compromised, biometric traits—such as fingerprints or facial features—are permanent and unique. Once an individual's biometric data is compromised, it cannot be reset or changed.
As AI technologies become more adept at replicating biometric traits, attackers are finding it easier to spoof security systems that were once considered highly reliable. The sophistication of multi-modal AI means that attackers no longer need access to a single biometric data point; they can instead leverage a combination of video, audio, and textual information to create a full profile of a target. This makes traditional authentication methods increasingly vulnerable, especially in high-risk sectors like banking and government services, where security is paramount.
Implications for Financial Services
In financial services, the stakes are particularly high. As digital identities become more integrated into online banking, digital wallets, and payment systems, the attack surface for malicious actors expands significantly. Multi-modal AI allows cybercriminals to craft more convincing and nuanced impersonations of customers or employees, potentially leading to financial fraud, data breaches, or even systemic risks. This poses a grave challenge for financial institutions, which must find ways to bolster the security of their identity verification systems while also maintaining a seamless user experience.
Accordingly, use of multi-modal AI in financial services could lead to a significant erosion of trust in digital identity systems. If individuals and organizations can no longer trust that their biometric data is safe, they may hesitate to adopt or fully integrate these technologies, potentially disrupting the growth of digital economies.
The Risks of Agentic AI
The Anthropic report highlights the significant escalation in AI attacks: “This campaign demonstrated unprecedented integration and autonomy of AI throughout the attack lifecycle, with the threat actor manipulating Claude Code to support reconnaissance, vulnerability discovery, exploitation, lateral movement, credential harvesting, data analysis, and exfiltration operations largely autonomously. The human operator tasked instances of Claude Code to operate in groups as autonomous penetration testing orchestrators and agents, with the threat actor able to leverage AI to execute 80-90% of tactical operations independently at physically impossible request rates.” The risks to the wider business community are clear, as highlighted by the attack recently featured on the front page of the Wall Street Journal.
Mitigating the Risk of Multi-Modal AI and Agentic AI
Cybersecurity compliance (including under HIPAA, the Gramm Leach Bliley Act, the FTC Safeguards Rule, and state laws such as the NY SHIELD Act, NYSDFS Cybersecurity Regulation, and Illinois Personal Information Protection Act) requires organizations to consider the risks from emerging cybersecurity threats and to implement reasonable risk based safeguards. As the above discussion indicates, these threats are not hypothetical, and as the threat landscape evolves, the need for more robust authentication and defensive systems becomes increasingly urgent. Organizations must look beyond traditional authentication methods and adopt multi-layered strategies to defend against AI-powered attacks. Some key strategies to mitigate the risks posed by multi-modal AI and Agentic AI include:
- AI Governance: Maintain an AI System governance, compliance and internal audit program that conducts risk assessments and considers the cyber threats in connection with the use of AI, including addressing threats highlighted by the Open Worldwide Application Security Project (OWASP) (such as Agentic AI, LLMs) and the National Institute of Standards and Technology (NIST).
- Multi-Factor Authentication: Combining several forms of authentication—such as biometrics, passwords, and device-based security—can provide a much stronger defense against AI-driven attacks. Adding multiple layers of protection significantly complicates the process for attackers trying to spoof a user's identity.
- Behavioral Biometrics: Moving beyond static biometrics, such as fingerprints or facial recognition, behavioral biometrics monitors user behavior patterns, such as how they type or interact with a device. These dynamic identifiers are much harder to replicate, making them a useful complement to traditional biometric systems.
- Continuous Authentication: Instead of relying on a one-time authentication process, continuous authentication tracks user behavior in real time. By monitoring factors like typing speed, mouse movements, or even voice inflections during a session, systems can detect anomalies that indicate fraud.
- AI Detection Tools: As deepfakes become more sophisticated, financial institutions and other organizations can invest in AI systems designed to detect synthetic media. These tools can identify inconsistencies or abnormalities in audio, video, and text data that are often present in AI-generated content.
- User Education and Awareness: Organizations must educate their users on the potential risks of multi-modal AI, especially as it relates to identity theft and fraud. Awareness programs can help users recognize phishing attempts and other forms of social engineering that exploit AI-driven impersonations.
- Secure APIs: Inventory and protect APIs.
Conclusion
The rapid development of multi-modal AI and agentic AI represents both an opportunity and a threat in the world of cybersecurity. While these technologies offer exciting advancements in fields like customer service, healthcare, manufacturing and finance, they also present a new class of risks—particularly in the realm of authentication and identity verification, ransomware and data theft. As cybercriminals become more adept at leveraging AI to bypass security systems, it is crucial that organizations and individuals adopt a multi-faceted approach to digital identity security. With the right safeguards in place, the promise of AI can be harnessed while mitigating its associated risks.
Blog Editors
Authors
- Member of the Firm
- Member of the Firm