Connect with us

Hi, what are you looking for?

AI Cybersecurity

AI-Driven Cyber Threats Surge: India Faces ₹20,000 Crore in Deepfake Fraud Losses

AI-driven deepfake fraud threatens India’s financial sector, with losses expected to exceed ₹20,000 crore by 2025, highlighting urgent cybersecurity needs.

AI-driven deepfake fraud threatens India's financial sector, with losses expected to exceed ₹20,000 crore by 2025, highlighting urgent cybersecurity needs.

In July 2025, a bank executive in India executed an urgent transfer of ₹1.08 crore after what he believed was a video call with his CEO. The video was flawless, the voice unmistakable, but it was all an elaborate deepfake. This incident underscores a troubling new frontier in cyber risk where artificial intelligence (AI) serves as both a tool for fraud detection and a means of perpetrating sophisticated scams.

The rise of AI-driven threats has become a pressing concern for India’s financial sector. The India Cyber Threat Report 2025 forecasts a steep increase in deepfake-enabled cyberattacks, particularly within the Banking, Financial Services, and Insurance (BFSI) sector. Estimates suggest that AI-enabled financial scams could lead to losses exceeding ₹20,000 crore by the close of 2025.

As AI technology advances, it has fundamentally altered the landscape of cybercrime. What once required advanced technical expertise can now be executed by anyone with access to readily available AI tools. Cybercriminals are exploiting this technology to create hyper-realistic videos and audio that impersonate executives, government officials, and trusted advisors. For instance, in 2024, fraudsters in Hong Kong utilized deepfake technology to impersonate a company’s Chief Financial Officer, ultimately swindling the firm out of $25 million.

In India, deepfake videos featuring Finance Minister Nirmala Sitharaman and Google CEO Sundar Pichai were deployed to promote fraudulent investment schemes, resulting in significant losses for numerous investors. Even cybersecurity experts find it challenging to detect these deepfakes in real-time, which exacerbates the problem.

Phishing attacks have also evolved with the introduction of AI. Traditional phishing schemes, often characterized by poor grammar and generic greetings, have been transformed. AI can now generate highly personalized phishing emails that emulate an organization’s communication style and reference recent interactions. Reports from CERT-In reveal that phishing now constitutes 38% of all reported fintech frauds in India, with attackers masquerading as the Reserve Bank of India and other trusted entities.

Another emerging threat is synthetic identity fraud, where AI generates entirely fictitious identities by blending real and fabricated information. Research from NASSCOM-DSCI indicates that the prevalence of fake identities has surged by 450% since 2022. These synthetic identities can evade Know Your Customer (KYC) checks, establish credit histories, and operate undetected for extended periods, often vanishing with borrowed funds.

AI also enhances Business Email Compromise (BEC) attacks, enabling criminals to monitor email patterns and organizational hierarchies. A deepfake audio clip mimicking a senior executive’s voice, combined with an urgent payment request, creates a scenario where even well-trained employees may be deceived.

The rapid pace and scale of AI-driven attacks have rendered traditional security measures inadequate. Legacy systems are ill-equipped to handle threats that adapt in real-time, and rule-based security tools cannot keep up with AI that learns from each encounter. Human verification processes falter when the individual being verified may be an AI-generated deepfake.

According to the Reserve Bank of India’s 2023-24 Currency and Finance report, cybersecurity threats, legacy systems, and low digital literacy among customers significantly impede digital adoption. However, this report was compiled before the full extent of AI-powered threats became apparent.

Addressing these challenges requires a shift in governance rather than merely a focus on technology. The Economic Survey 2025 indicates that one in five reported cyberattacks in India targeted the BFSI sector in 2024, and this ratio is anticipated to worsen as AI-driven attacks grow in sophistication.

Boards of directors must take cybersecurity seriously and cannot delegate it solely to IT departments. They need to understand the AI attack surface across their organization, establish protocols for verifying authenticity during critical communications, ensure third-party vendor security, and prepare incident response plans that consider AI-driven attacks.

Utilizing AI responsibly is essential for both defense and differentiation in this new landscape. AI-based fraud detection systems can identify subtle anomalies that humans might overlook. Implementing behavioral biometrics and multi-signal verification further strengthens defenses against sophisticated attacks.

As regulatory bodies begin to respond, frameworks must be updated to address the unique challenges posed by AI-driven threats. The Reserve Bank of India’s guidelines from 2016 and the Securities and Exchange Board of India’s Cybersecurity and Cyber Resilience Framework require urgent revisions to include mandatory deepfake detection capabilities and AI model governance.

Organizations that integrate AI adoption with robust cybersecurity measures will find themselves at an advantage. By fostering a culture that prioritizes verification and implementing comprehensive incident response plans, they can navigate the complex threats of the AI age. The cost of unpreparedness is no longer just a data breach; it could mean substantial financial losses and irreparable damage to customer trust.

As Santanu Sengupta, a board leader and former Managing Director at Wells Fargo Bank, aptly notes, the question is not whether an organization will face AI-powered cyberattacks but whether it will be prepared when they arrive.

See also
Rachel Torres
Written By

At AIPressa, my work focuses on exploring the paradox of AI in cybersecurity: it's both our best defense and our greatest threat. I've closely followed how AI systems detect vulnerabilities in milliseconds while attackers simultaneously use them to create increasingly sophisticated malware. My approach: explaining technical complexities in an accessible way without losing the urgency of the topic. When I'm not researching the latest AI-driven threats, I'm probably testing security tools or reading about the next attack vector keeping CISOs awake at night.

You May Also Like

AI Finance

TRON integrates B.AI into its blockchain, enabling $22 billion daily automated payments, revolutionizing financial infrastructure for AI-driven economies.

AI Regulation

Butterfly Data calls for public sector organizations to prioritize data provenance in AI development, highlighting that data origins are crucial for fairness and compliance.

Top Stories

Google launches the Gemini app for Mac, its first native macOS AI assistant, enhancing desktop access with customizable shortcuts and screen sharing features.

AI Marketing

Emplifi's new report reveals 93% of consumers believe authentic engagement fosters trust, highlighting the critical role of transparency in AI-driven marketing.

AI Generative

OpenAI is set to launch GPT-6 this week, featuring significant upgrades like a larger context window and native multimodal capabilities.

Top Stories

Rent the Runway announces a transformative AI-driven strategy for 2026, focusing on personalized fashion discovery to enhance subscriber engagement after achieving $329.8M in revenue.

AI Regulation

Illinois lawmakers engage in virtual hearings on 50 AI bills, addressing urgent consumer protection concerns highlighted by Senator Mary Edly-Allen's warnings against past tech...

Top Stories

Mistral unveils Connectors in Studio, enabling seamless API integration for enterprise AI applications, streamlining workflows and reducing setup time significantly.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.