Your vision will become clear only when you look into your heart.... Who looks outside, dreams. Who looks inside, awakens. Carl Jung
Sunday, November 16, 2025
AI Deepfakes
https://www.citigroup.com/rcs/citigpa/storage/public/Citi_Institute_Report_AI_Deepfakes.pdf
AI Deepfakes
Key Takeaways
AI Deepfakes | When Seeing and Hearing Can’t be Trusted
© 2025 Citigroup3
AI Deepfakes
Deepfakes are the new face of deception.They are voices, images, videos or even text created by artificial intelligence (AI) that look and sound indistinguishably real. Once mere entertainment novelties, they have evolved into powerful tools of manipulation and fraud, marking a new era in financial crime. This wave of AI-driven deception is now infiltrating the workplace and recruitment processes. Some estimates suggest that by 2028, one in four candidate profiles worldwide could be fake.4 Meanwhile, one technology company the Citi Institute spoke with at Money 20/20 USA (October 2025) told us that 50% of job applications it receives are fake. Identity deception is accelerating. AIimprovementsenable deepfakes to mimic real people and make it difficult to detect synthetic ones. The rise of
video-based hiring and remote work, especially in the technology and web3 sectors, further amplifies the risk. As recruitment interviews shift to virtual environments, imposters with fake credentials may be able to secure jobs. The danger lies not just in what they fake, but in how long they can stay undetected, infiltrating sensitive systems and installing malware or ransomware.
State-sponsored Deepfake Employees
A striking example is the surge innstate-sponsored actors using deepfake technology to infiltrate global companies. These candidates are either completely AI-generated for have their appearance significantly altered using deepfake technology. North Korea has emerged as a hotspot,with its operatives often posing as IT professionals aiming to infiltrate foreign
companies using false identities. Anestimated 320 companies have been infiltrated by North Korean IT workers in the past year.5,6 The deepfake employee scam tends to be
a long-term campaign involving multiple individuals across locations including
overseas operators (impersonators),onshore collaborators (mules), brokers,and money movement handlers. A typical operation includes:
• Targeting remote roles with limited in-person onboarding
• Creating synthetic identities with fabricated CVs, social profiles, and
realistic headshots or videos
• Participating in deepfake interviews using voice or video cloning, providing
fabricated work history and personal references
• Establishing an onshore foothold through a rented address or a local
collaborator for administrative purposes and deliveries
• Having company laptops or authenticators mailed to the
onshore address
• Connecting remotely to company systems using the provided credentials and hardware, often with the help from the collaborator
• Co-ordinating multi-location teams to scale the operation. Once candidates gain access to the company ecosystem, the scam goes beyond simple deception. Potential impacts include the infiltration of sensitive systems and leakage of trade secrets, theft of intellectual property and customer data, data breaches and ransomware attacks, generation of foreign currency revenues for the sponsoring regime, and the erosion
of trust and reputation. For one job posting alone, we received over 800 applications in a matter of days.
“When we conducted a deeper analysis of 300 candidate profiles, over one-third were outright fraudulent. These weren’t just candidates exaggerating their experience –
these were entirely fabricated identities, many leveraging AI-generated resumes,
manipulated credentials, and, most concerning, deepfake video interviews.” Vijay Balasubramaniyan, Co-Founder and CEO, Pindrop Security
AI Deepfakes | When Seeing and Hearing Can’t be Trusted
© 2025 Citigroup4
This is not just a problem for large corporates. Small and mid-sized businesses, which often lack the resources to detect sophisticated hiring fraud, are particularly vulnerable.Corporates often fail to conductthorough verification of remote hires and contractors. Standard background checks often rely on self-reported information or basic identity verification,which can be manipulated through deepfake visuals, fabricated credentials,or stolen digital identities.Deepfakes are not confined to job applicants. They also impact senior leadership, customers, and suppliers. In the financial sector, this could extend to synthetic identity creation, fraudulent transaction authorization, and automated money transfer scams.
Financial Deepfake Fraud
There have been several single-event,high-impact financial frauds where deepfakes impersonated trusted executives to authorize or redirect payments. Unlike systemic infiltration cases, the objective
here is immediate financial gain rather than long-term access.
In a widely reported case in 2024, a UK multinational became the targetof a stri kingly sophisticated fraud.An employee of its Hong Kong office received what appeared to be a video callfrom the chief financial officer and other senior leaders, urging urgent money transfers for a confidential transaction.7
In reality, the voices and images of the executives had been deepfaked and every participant on the call except the employee was a synthetic representation. The employee proceeded to make 15 separate transfers totalling around HKD200 million ($25 million) to an offshore account before the deception was uncovered.
Several such incidences have been reported in recent years, with potentially more going unreported. Such incidents illustrate the danger of deepfakes,rendering traditional verification controls such as voice recognition and visual confirmation inadequate. It also highlights the importance of continuous education to understand the common techniques that fraudsters use to enable the early detection of fakes.More importantly, ongoing and proactive education is key for law enforcement to stay ahead of the perpetrators.
Audio Spoofs to Full-Motion
Real Time Video
Deepfake fraud attempts are multiplying fast. 2024 saw a marked increase indeep fake-related fraud, accounting for nearly 4.7% of all fraud attempts,up from 0.10% in 2022.8.Deepfake fraud attempts vary by sector:
in consumer credit they represent nearly
12% of all fraud, followed by real estate
(about 8%) and payments (5%). By
contrast, fewer deepfake attempts
occur where the perceived gains are
lower, such as education.9
Up to 8 million deepfakes are expected
to be shared online by the end of 2025,
up from 500,000 in 2023, suggesting a
doubling every six months.10 Easy access
to powerful AI tools and vast amounts
of data are likely contributing factors.
A survey of over 100 fraud executives
from global financial institutions over
2Q2025-3Q2025 suggest greater
anticipation of increased fraud losses
in banking payments over the next
three years, especially for the U.S.11
The increase is attributed to growth
in AI-powered deepfake attacks
and synthetic identity fraud.
Source: Signicat VideoID data, full-year 2021–2024, from “The Battle in the Dark”
2022
0.10%
2023
2.52%
2024
4.70%
Figure 1. Deepfake fraud as percentage of all fraud attempts
AI Deepfakes | When Seeing and Hearing Can’t be Trusted
© 2025 Citigroup5
Voice deepfakes: Voice deepfakes are
synthetic audio created to sound like
the target individual. They are created
by training a model on samples of a
person’s voice and then providing a
text to generate the fake speech.
Combining large language models (LLMs)
with text-to-speech engines can enable
voice bots to respond in real time. The
LLM generates real-time answers to
questions, and the text-to-speech
engine vocalizes them, even conveying
emotions such as empathy or urgency.
Video deepfakes: These extend the
threat to full visual manipulation.
Initially, video deepfakes were
pre-recorded content that was
manipulated before distribution.
Now advanced deepfakes can manipulate
someone’s likeness using generative
adversarial networks and other machine
learning techniques designed to create
lifelike synthetic media.
Attackers often use publicly available
data like photos, videos, and audio from
social media websites or corporate filings
to build the model. During a live call, the
model overlays the synthetic likeness in
real time, replicating facial cues, voice,
and gestures.
Hybrid deepfakes: Many live deepfake
scams use a hybrid strategy, combining
multiple types of deepfakes (including
voice and video), with traditional
social engineering tactics, fabricated
documents, and credential theft.
Instead of a one-off scam, such
deepfakes are often associated with
long-term planned infiltration campaigns.
They are designed to build trust, before
compromising systems or extracting
sensitive data.
Detecting Deepfakes is
Getting Harder
Detecting deepfakes is getting harder,
especially audio. Early versions often
contained noticeable pauses as the
operator typed responses. Recent
iterations eliminate these flaws,
producing seamless and natural-
sounding speech.
The number of generative AI (GenAI)
systems capable of cloning voice and
video has surged, rising from roughly
100-150 tracked systems last year to
more than 500 today.13 Much of this
growth is driven by open-source tools
that are easier to access, cheaper
to run, and require less data to
create highly convincing fakes.
There are cases of bots simulating
empathy. For example, a bot may
remark during an interaction “It must
have been a long day. So how are you
holding up?” This carefully engineered
scripted empathy, helps increase
credibility and makes the interaction
appear more convincing.
In contrast, video deepfakes still
show some tell-tale flaws such as
blurriness or unnatural pixelation.
But the technology is evolving fast.
Figure 2. Key forces behind the growing speed and sophistication of speech-based deepfakes12
Use of automated bots
Previously, speech generation tools
had a 4-7 second delay between input
and synthetic voice output. Today,
LLMs have reduced that delay to near
real-time. This makes it increasingly
difficult to distinguish synthetic
voices from real ones.
Emotional-sounding AI
Advances in synthetic speech
have enabled text-to-speech
voices to convey emotions like joy,
anger, empathy, and sadness. AI
models can now learn and imitate
emotional tones from human
speech, making these synthetic
voices even more convincing.
Real-time voice conversion
Companies have created tools for
real-time voice conversion, allowing
users to change pitch, timbre, and
accent instantly. While this technology
benefits voice dubbing, gaming, and
content creation, it also makes it
easier for fraudsters to evade voice
recognition systems by masking
their voice.
Source: Pindrop 2025 Voice Intelligence and Security Report, Citi Institute
6 © 2025 Citigroup
AI Deepfakes | When Seeing and Hearing Can’t be Trusted
How are Corporates Responding?
The global annual cost of cybercrime is
estimated to reach $10.5 trillion by 2025,
up from $3 trillion in 2015.14 Old school
fraud using one-time passwords (OTPs)
and phishing continue to exist, but
deepfakes are seeing a rapid increase.
Corporates are responding in different
ways. Many firms are prioritizing the
detection of C-suite impersonation
due to the high-profile nature of these
attacks targeting CEOs and CFOs. Others
are investing in tools to safeguard video
communications more broadly.
As fraud grows more sophisticated,
traditional identity checks such as
document scans or liveness tests are
no longer sufficient. Trust cannot be
established through a single interaction.
Verification must evolve into
multi-layered digital constructs that
combine biometrics, behavior and
device data, and contextual cues.
The notion of continuous identity
is becoming crucial.
Zero-Trust Communication
is Essential
In a GenAI-powered world, trust can
no longer be assumed, it must be
continuously verified.
Every interaction, whether from inside
or outside the network, must be verified
through multiple layers of identity,
device, and behavioural validation.
The principle must be “Never trust,
always verify”.
While corporates are redesigning call
centres for zero trust, communication
channels like phone calls and emails
remain dangerously outdated. Many
corporate systems still rely on voice
recognition, caller ID, and email
domain as proof of authenticity.
The next frontier is zero-trust
communication, where every
conversation and message undergo
real-time authentication using biometric
voiceprints, behavioral analytics, and
device-level identity tokens. Likewise,
email security must move towards
cryptographic message signing,
AI-based anomaly detection,
and intent verification.
Fighting AI with AI
The way financial services combat fraud
will fundamentally change as criminals
adopt AI to perpetrate scams. Deepfakes’
ability to circumvent traditional defenses
illustrate this shift. While the pace of
AI-driven fraud is alarming, the fight
against deepfakes is winnable.
The same AI technology that enables
fraud, can also be used against it.
Advanced AI agents are now capable
of mapping scam networks, flagging
manipulated audio and video, and
intercepting social engineering
attempts with increasing precision.
Building AI-driven defense systems
is becoming as critical to financial
security as cybersecurity firewalls.
As AI agents evolve and operate
autonomously, the risks escalate.
Bad actors can deploy agents at scale
to impersonate senior executives,
manipulate employees, or mislead
customers. This raises the bar for
verification. Financial institutions
must move beyond validating users
to also verifying the identity, intent,
and provenance of AI agents.
In Citi GPS: Agentic AI we highlight
several examples of how AI is being
used to counter fraud. One leading global
bank, for instance, integrated real-time
deepfake detection into its call center
infrastructure. The detection process
happens seamlessly without introducing
latency or disrupting the natural flow
of conversation. The AI tool analyses
the audio stream and flags signs of
synthetic manipulation.
The industry is also beginning to build
frameworks such as Know Your Agent
(KYA), mirroring the KYC standard, to
safeguard trust in digital interactions.
AI Deepfakes | When Seeing and Hearing Can’t be Trusted
© 2025 Citigroup7
Endnotes
1 Gartner, Gartner Survey Shows Just 26% of Job Applicants Trust AI Will Fairly Evaluate Them,
31 July 2025.
2 Fortune, North Korean IT Worker Infiltrations Exploded 220% Over the Past 12 months, with GenAI
Weaponized at Every Stage of the Hiring Process, 04 August 2025; Crowdstrike, Threat Hunting
Report, 2025.
3 eSentire, Cybercrime to Cost the World $9.5 Trillion USD Annually in 2024.
4 Gartner, Gartner Survey Shows Just 26% of Job Applicants Trust AI Will Fairly Evaluate Them,
31 July 2025.
5 Fortune, North Korean IT Worker Infiltrations Exploded 220% Over the Past 12 months, with
GenAI Weaponized at Every Stage of the Hiring Process, 04 August 2025.
6 Crowdstrike, Threat Hunting Report, 2025.
7 CNN Business, British Engineering Giant Arup Revealed as $25 million Deepfake Scam Victim,
17 May 2024.
8 Signicat, The Battle in the Dark, October 2025.
9 Signicat, The Battle in the Dark, October 2025.
10 UK Government (UK.Gov), Innovating to Detect Deepfakes and Protect
the Public, 05 February 2025.
11 Datos Insights, Five Forces Disrupting Global Fraud Prevention by 2030, October 2025.
12 Pindrop 2025 Voice Intelligence and Security Report.
13 Citi Institute Future of Finance Forum 2025 Video, Deep Dive into Deepfakes, 09 July 2025.
14 eSentire, Cybercrime to Cost the World $9.5 Trillion USD Annually in 2024.
Authors
Contributors
Vijay Balasubramaniyan
Pindrop Security
Sophia Bantanidis
Future of Finance,
Citi Institute
sophia.bantanidis@citi.com
Kaiwan Master
Future of Finance,
Citi Institute
kaiwan.hoshang.master@citi.com
Ronak Shah
Future of Finance,
Citi Institute
ronak.sharad.shah@citi.com
Prag Sharma
Future of Finance,
Citi Institute
prag.sharma@citi.com
Ronit Ghose
Global Head, Future
of Finance, Citi Institute
ronit.ghose@citi.com