Sunday, November 16, 2025

AI Deepfakes

https://www.citigroup.com/rcs/citigpa/storage/public/Citi_Institute_Report_AI_Deepfakes.pdf AI Deepfakes Key Takeaways AI Deepfakes | When Seeing and Hearing Can’t be Trusted © 2025 Citigroup3 AI Deepfakes Deepfakes are the new face of deception.They are voices, images, videos or even text created by artificial intelligence (AI) that look and sound indistinguishably real. Once mere entertainment novelties, they have evolved into powerful tools of manipulation and fraud, marking a new era in financial crime. This wave of AI-driven deception is now infiltrating the workplace and recruitment processes. Some estimates suggest that by 2028, one in four candidate profiles worldwide could be fake.4 Meanwhile, one technology company the Citi Institute spoke with at Money 20/20 USA (October 2025) told us that 50% of job applications it receives are fake. Identity deception is accelerating. AIimprovementsenable deepfakes to mimic real people and make it difficult to detect synthetic ones. The rise of video-based hiring and remote work, especially in the technology and web3 sectors, further amplifies the risk. As recruitment interviews shift to virtual environments, imposters with fake credentials may be able to secure jobs. The danger lies not just in what they fake, but in how long they can stay undetected, infiltrating sensitive systems and installing malware or ransomware. State-sponsored Deepfake Employees A striking example is the surge innstate-sponsored actors using deepfake technology to infiltrate global companies. These candidates are either completely AI-generated for have their appearance significantly altered using deepfake technology. North Korea has emerged as a hotspot,with its operatives often posing as IT professionals aiming to infiltrate foreign companies using false identities. Anestimated 320 companies have been infiltrated by North Korean IT workers in the past year.5,6 The deepfake employee scam tends to be a long-term campaign involving multiple individuals across locations including overseas operators (impersonators),onshore collaborators (mules), brokers,and money movement handlers. A typical operation includes: • Targeting remote roles with limited in-person onboarding • Creating synthetic identities with fabricated CVs, social profiles, and realistic headshots or videos • Participating in deepfake interviews using voice or video cloning, providing fabricated work history and personal references • Establishing an onshore foothold through a rented address or a local collaborator for administrative purposes and deliveries • Having company laptops or authenticators mailed to the onshore address • Connecting remotely to company systems using the provided credentials and hardware, often with the help from the collaborator • Co-ordinating multi-location teams to scale the operation. Once candidates gain access to the company ecosystem, the scam goes beyond simple deception. Potential impacts include the infiltration of sensitive systems and leakage of trade secrets, theft of intellectual property and customer data, data breaches and ransomware attacks, generation of foreign currency revenues for the sponsoring regime, and the erosion of trust and reputation. For one job posting alone, we received over 800 applications in a matter of days. “When we conducted a deeper analysis of 300 candidate profiles, over one-third were outright fraudulent. These weren’t just candidates exaggerating their experience – these were entirely fabricated identities, many leveraging AI-generated resumes, manipulated credentials, and, most concerning, deepfake video interviews.” Vijay Balasubramaniyan, Co-Founder and CEO, Pindrop Security AI Deepfakes | When Seeing and Hearing Can’t be Trusted © 2025 Citigroup4 This is not just a problem for large corporates. Small and mid-sized businesses, which often lack the resources to detect sophisticated hiring fraud, are particularly vulnerable.Corporates often fail to conductthorough verification of remote hires and contractors. Standard background checks often rely on self-reported information or basic identity verification,which can be manipulated through deepfake visuals, fabricated credentials,or stolen digital identities.Deepfakes are not confined to job applicants. They also impact senior leadership, customers, and suppliers. In the financial sector, this could extend to synthetic identity creation, fraudulent transaction authorization, and automated money transfer scams. Financial Deepfake Fraud There have been several single-event,high-impact financial frauds where deepfakes impersonated trusted executives to authorize or redirect payments. Unlike systemic infiltration cases, the objective here is immediate financial gain rather than long-term access. In a widely reported case in 2024, a UK multinational became the targetof a stri kingly sophisticated fraud.An employee of its Hong Kong office received what appeared to be a video callfrom the chief financial officer and other senior leaders, urging urgent money transfers for a confidential transaction.7 In reality, the voices and images of the executives had been deepfaked and every participant on the call except the employee was a synthetic representation. The employee proceeded to make 15 separate transfers totalling around HKD200 million ($25 million) to an offshore account before the deception was uncovered. Several such incidences have been reported in recent years, with potentially more going unreported. Such incidents illustrate the danger of deepfakes,rendering traditional verification controls such as voice recognition and visual confirmation inadequate. It also highlights the importance of continuous education to understand the common techniques that fraudsters use to enable the early detection of fakes.More importantly, ongoing and proactive education is key for law enforcement to stay ahead of the perpetrators. Audio Spoofs to Full-Motion Real Time Video Deepfake fraud attempts are multiplying fast. 2024 saw a marked increase indeep fake-related fraud, accounting for nearly 4.7% of all fraud attempts,up from 0.10% in 2022.8.Deepfake fraud attempts vary by sector: in consumer credit they represent nearly 12% of all fraud, followed by real estate (about 8%) and payments (5%). By contrast, fewer deepfake attempts occur where the perceived gains are lower, such as education.9 Up to 8 million deepfakes are expected to be shared online by the end of 2025, up from 500,000 in 2023, suggesting a doubling every six months.10 Easy access to powerful AI tools and vast amounts of data are likely contributing factors. A survey of over 100 fraud executives from global financial institutions over 2Q2025-3Q2025 suggest greater anticipation of increased fraud losses in banking payments over the next three years, especially for the U.S.11 The increase is attributed to growth in AI-powered deepfake attacks and synthetic identity fraud. Source: Signicat VideoID data, full-year 2021–2024, from “The Battle in the Dark” 2022 0.10% 2023 2.52% 2024 4.70% Figure 1. Deepfake fraud as percentage of all fraud attempts AI Deepfakes | When Seeing and Hearing Can’t be Trusted © 2025 Citigroup5 Voice deepfakes: Voice deepfakes are synthetic audio created to sound like the target individual. They are created by training a model on samples of a person’s voice and then providing a text to generate the fake speech. Combining large language models (LLMs) with text-to-speech engines can enable voice bots to respond in real time. The LLM generates real-time answers to questions, and the text-to-speech engine vocalizes them, even conveying emotions such as empathy or urgency. Video deepfakes: These extend the threat to full visual manipulation. Initially, video deepfakes were pre-recorded content that was manipulated before distribution. Now advanced deepfakes can manipulate someone’s likeness using generative adversarial networks and other machine learning techniques designed to create lifelike synthetic media. Attackers often use publicly available data like photos, videos, and audio from social media websites or corporate filings to build the model. During a live call, the model overlays the synthetic likeness in real time, replicating facial cues, voice, and gestures. Hybrid deepfakes: Many live deepfake scams use a hybrid strategy, combining multiple types of deepfakes (including voice and video), with traditional social engineering tactics, fabricated documents, and credential theft. Instead of a one-off scam, such deepfakes are often associated with long-term planned infiltration campaigns. They are designed to build trust, before compromising systems or extracting sensitive data. Detecting Deepfakes is Getting Harder Detecting deepfakes is getting harder, especially audio. Early versions often contained noticeable pauses as the operator typed responses. Recent iterations eliminate these flaws, producing seamless and natural- sounding speech. The number of generative AI (GenAI) systems capable of cloning voice and video has surged, rising from roughly 100-150 tracked systems last year to more than 500 today.13 Much of this growth is driven by open-source tools that are easier to access, cheaper to run, and require less data to create highly convincing fakes. There are cases of bots simulating empathy. For example, a bot may remark during an interaction “It must have been a long day. So how are you holding up?” This carefully engineered scripted empathy, helps increase credibility and makes the interaction appear more convincing. In contrast, video deepfakes still show some tell-tale flaws such as blurriness or unnatural pixelation. But the technology is evolving fast. Figure 2. Key forces behind the growing speed and sophistication of speech-based deepfakes12 Use of automated bots Previously, speech generation tools had a 4-7 second delay between input and synthetic voice output. Today, LLMs have reduced that delay to near real-time. This makes it increasingly difficult to distinguish synthetic voices from real ones. Emotional-sounding AI Advances in synthetic speech have enabled text-to-speech voices to convey emotions like joy, anger, empathy, and sadness. AI models can now learn and imitate emotional tones from human speech, making these synthetic voices even more convincing. Real-time voice conversion Companies have created tools for real-time voice conversion, allowing users to change pitch, timbre, and accent instantly. While this technology benefits voice dubbing, gaming, and content creation, it also makes it easier for fraudsters to evade voice recognition systems by masking their voice. Source: Pindrop 2025 Voice Intelligence and Security Report, Citi Institute 6 © 2025 Citigroup AI Deepfakes | When Seeing and Hearing Can’t be Trusted How are Corporates Responding? The global annual cost of cybercrime is estimated to reach $10.5 trillion by 2025, up from $3 trillion in 2015.14 Old school fraud using one-time passwords (OTPs) and phishing continue to exist, but deepfakes are seeing a rapid increase. Corporates are responding in different ways. Many firms are prioritizing the detection of C-suite impersonation due to the high-profile nature of these attacks targeting CEOs and CFOs. Others are investing in tools to safeguard video communications more broadly. As fraud grows more sophisticated, traditional identity checks such as document scans or liveness tests are no longer sufficient. Trust cannot be established through a single interaction. Verification must evolve into multi-layered digital constructs that combine biometrics, behavior and device data, and contextual cues. The notion of continuous identity is becoming crucial. Zero-Trust Communication is Essential In a GenAI-powered world, trust can no longer be assumed, it must be continuously verified. Every interaction, whether from inside or outside the network, must be verified through multiple layers of identity, device, and behavioural validation. The principle must be “Never trust, always verify”. While corporates are redesigning call centres for zero trust, communication channels like phone calls and emails remain dangerously outdated. Many corporate systems still rely on voice recognition, caller ID, and email domain as proof of authenticity. The next frontier is zero-trust communication, where every conversation and message undergo real-time authentication using biometric voiceprints, behavioral analytics, and device-level identity tokens. Likewise, email security must move towards cryptographic message signing, AI-based anomaly detection, and intent verification. Fighting AI with AI The way financial services combat fraud will fundamentally change as criminals adopt AI to perpetrate scams. Deepfakes’ ability to circumvent traditional defenses illustrate this shift. While the pace of AI-driven fraud is alarming, the fight against deepfakes is winnable. The same AI technology that enables fraud, can also be used against it. Advanced AI agents are now capable of mapping scam networks, flagging manipulated audio and video, and intercepting social engineering attempts with increasing precision. Building AI-driven defense systems is becoming as critical to financial security as cybersecurity firewalls. As AI agents evolve and operate autonomously, the risks escalate. Bad actors can deploy agents at scale to impersonate senior executives, manipulate employees, or mislead customers. This raises the bar for verification. Financial institutions must move beyond validating users to also verifying the identity, intent, and provenance of AI agents. In Citi GPS: Agentic AI we highlight several examples of how AI is being used to counter fraud. One leading global bank, for instance, integrated real-time deepfake detection into its call center infrastructure. The detection process happens seamlessly without introducing latency or disrupting the natural flow of conversation. The AI tool analyses the audio stream and flags signs of synthetic manipulation. The industry is also beginning to build frameworks such as Know Your Agent (KYA), mirroring the KYC standard, to safeguard trust in digital interactions. AI Deepfakes | When Seeing and Hearing Can’t be Trusted © 2025 Citigroup7 Endnotes 1 Gartner, Gartner Survey Shows Just 26% of Job Applicants Trust AI Will Fairly Evaluate Them, 31 July 2025. 2 Fortune, North Korean IT Worker Infiltrations Exploded 220% Over the Past 12 months, with GenAI Weaponized at Every Stage of the Hiring Process, 04 August 2025; Crowdstrike, Threat Hunting Report, 2025. 3 eSentire, Cybercrime to Cost the World $9.5 Trillion USD Annually in 2024. 4 Gartner, Gartner Survey Shows Just 26% of Job Applicants Trust AI Will Fairly Evaluate Them, 31 July 2025. 5 Fortune, North Korean IT Worker Infiltrations Exploded 220% Over the Past 12 months, with GenAI Weaponized at Every Stage of the Hiring Process, 04 August 2025. 6 Crowdstrike, Threat Hunting Report, 2025. 7 CNN Business, British Engineering Giant Arup Revealed as $25 million Deepfake Scam Victim, 17 May 2024. 8 Signicat, The Battle in the Dark, October 2025. 9 Signicat, The Battle in the Dark, October 2025. 10 UK Government (UK.Gov), Innovating to Detect Deepfakes and Protect the Public, 05 February 2025. 11 Datos Insights, Five Forces Disrupting Global Fraud Prevention by 2030, October 2025. 12 Pindrop 2025 Voice Intelligence and Security Report. 13 Citi Institute Future of Finance Forum 2025 Video, Deep Dive into Deepfakes, 09 July 2025. 14 eSentire, Cybercrime to Cost the World $9.5 Trillion USD Annually in 2024. Authors Contributors Vijay Balasubramaniyan Pindrop Security Sophia Bantanidis Future of Finance, Citi Institute sophia.bantanidis@citi.com Kaiwan Master Future of Finance, Citi Institute kaiwan.hoshang.master@citi.com Ronak Shah Future of Finance, Citi Institute ronak.sharad.shah@citi.com Prag Sharma Future of Finance, Citi Institute prag.sharma@citi.com Ronit Ghose Global Head, Future of Finance, Citi Institute ronit.ghose@citi.com