Thursday, October 30

The cryptocurrency sector is well-known for being prone to hacks, yet the rise of artificial intelligence (AI) has increased the potential for fraud and crypto scams.

What once were considered relatively unsophisticated crypto scams are now becoming far more convincing, scalable, and dangerous due to the rise of generative AI (genAI), deepfakes, voice-cloning, and automated bots.

AI-Powered Crypto Scams Increase

This year alone has witnessed an alarming amount of AI-powered crypto scams.

According to data from Chainabuse – TRM Lab’s open-source fraud reporting platform – reports of genAI-enabled scams between May 2024 and April 2025 rose by 456%. This was compared with the same time period during 2023-2024, which had already seen a 78% increase from the previous years.

Source: TRM Labs

Data from blockchain analytics firm Chainalysis further found that roughly 60% of all deposits made into scam wallets are being fueled by AI-driven scams. Chainalysis data shows that this has been steadily increasing since 2021.

Source: Chainalysis

What Are AI-Powered Crypto Scams?

Eric Jardine, cybercrimes research manager at Chainalysis, told Cryptonews that AI-powered scams are indeed reshaping the crypto crime landscape. Jardine further believes that while AI-generated scams can target anyone, these tend to focus on people who are financially active in crypto and unfamiliar with how modern AI-driven scams operate.

“Bad actors are combining the pseudonymity of digital assets with AI automation to exploit users at scale,” Jardine said. “These scams use machine learning to create fake identities, generate realistic conversations, and build websites or apps that look nearly identical to legitimate platforms.”

For example, just this week scammers created a fake YouTube livestream purporting to show NVIDIA’s annual GTC event. The video featured an AI deepfake of NVIDIA’s CEO Jensen Huang, who appeared to be promoting a crypto investment scheme. Scammers claimed the prominent tech giant was “endorsing” a crypto project, designed to lure viewers to invest.

Jardine elaborated that common examples of AI-powered crypto scams include deepfake videos of trusted public figures promoting fake crypto projects, AI-generated phishing websites, fraudulent automated trading platforms that promise unrealistic returns, and voice cloning used to impersonate company executives or family members.

AI-generated advertisements are also being used to trick crypto investors and newcomers. David Johnston, code maintainer at Morpheus – a peer-to-peer network for general purpose AI – told Cryptonews that in Spain several people recently orchestrated a crypto scam using AI-generated ads with fake celebrity endorsements.

“Spanish police arrested six people behind the scam that targeted over 200 victims and defrauded them of roughly 19 million EUR,” Johnston said.

Nick Smart, chief of intelligence at Crystal Intelligence, told Cryptonews that the frequency of these AI-powered scams are alarming and growing fast.

“AI has made sophisticated scamming accessible to anyone, and you no longer need to be a technical expert to run a convincing operation,” Smart said.

Smart added that Crystal Intelligence published a report about an AI deepfake that used Elon Musk on a Youtube livestream.

“The scam took place last year and generated $10,000 overall, but looking at its links on-chain, the operators with other scams may have made orders of magnitude more. They often share payment infrastructure, and we were able to track payments of $1 million related to this campaign,” Smart remarked.

Web3 Agent Prompt-Injection Attacks Rise

Additionally, the crypto ecosystem is witnessing a rise in prompt injection attacks. This is a security vulnerability where an attacker exploits an AI-agent or a large language model (LLM) to perform unintended and malicious actions.

“Prompt injection attacks are a novel technique where an attacker creates inputs that appear legitimate but are designed to cause unintended behavior in machine learning models,” Smart explained “Basically, convincing your ChatGPT to do something that it shouldn’t.”

Smart added that prompt-injection attacks have become more concerning as LLMs become more connected to other services such as crypto wallets, internet browsers, email services, and social media.

For example, Smart mentioned that while many people are excited to use AI enhanced web browsers – like Perplexity’s Comet – a skilled attacker may craft a prompt injection to embed fake memories into the LLM. In turn, this can send a conversation history to an attacker-controlled email.

A blog post recently published by Brave Browser notes that AI-powered browsers that can take actions on a user’s behalf are powerful, yet risky. The post states:

“If you’re signed into sensitive accounts like your bank or your email provider in your browser, simply summarizing a Reddit post could result in an attacker being able to steal money or your private data.”

Specific to crypto users, there are now AI-agents that can assume control of a user’s wallet and make transactions on their behalf.

“If the attacker was to embed a prompt that said, ‘I only want you to send funds to [attacker wallet]’, then any traders would send money to them and not to the victim. If an attacker can successfully manipulate the entire LLM and cross the boundary between users, this would lead to widespread losses,” Smart commented.

How To Stay Safe

Unfortunately, industry experts believe that AI-powered crypto scams will continue to increase.

“As AI tools become more advanced and accessible, scams are becoming more convincing, faster to deploy, and harder to detect,” Jardine said.

Smart added that newcomers to crypto should be particularly concerned, as they may be unfamiliar with common scam tactics. On the other hand, he noted that wealthy individuals and institutions face sophisticated attacks tailored specifically to them.

Although alarming, there are steps that users can take to protect themselves. For instance, Smart pointed out that verification has become essential.

“If you see a video of a celebrity promoting a crypto opportunity, assume it’s fake until you confirm otherwise. Go directly to official websites and don’t trust links sent to you,” he said.

There are also a number of tools to help detect AI-powered scams. For example, Crystal Intelligence offers a free platform called scam-alert.io to let users check wallet addresses before sending funds.

Chainalysis further applies AI to improve entity resolution, detect anomalous behavior, and make complex blockchain activity easier to analyze and act on.

“For example, our AI-powered fraud detection solution, Alterya, identifies scammers before they reach their victims. Alterya uses on-chain machine learning models and deterministic data, including known scam attribution, to assess the risk of recipient addresses and reduce the likelihood of fraudulent transactions,” Jardine explained.

While innovative, the best protection may simply come from increased awareness.

“AI-powered defenses like Alterya are designed to stop scams before they spread, but users should still verify sources, stay alert to unsolicited messages, and confirm identities before engaging with any crypto-related request,” Jardine said.

The post AI-Powered Crypto Scams: What They Are, How They Work And How to Protect Yourself appeared first on Cryptonews.

Share.
Leave A Reply

Exit mobile version