Here’s how to quickly spot a deepfake crypto scam — cybersecurity execs

The fast-paced nature of the crypto markets means investors are under massive pressure to quickly verify whether a video message is authentic or not.

Crypto investors have been urged to keep their eyes peeled for “deepfake” crypto scams to come, with the digital-doppelganger technology continuing to advance, making it harder for viewers to separate fact from fiction. 

David Schwed, the chief operating officer of blockchain security firm Halborn, told Cointelegraph that the crypto industry is more “susceptible” to deepfakes than ever because “time is of the essence in making decisions,” which results in less time to verify the veracity of a video.

Cast your vote now!

Deepfakes use deep learning artificial intelligence (AI) to create highly realistic digital content by manipulating and altering original media, such as swapping faces in videos, photos, and audio, according to OpenZeppelin technical writer Vlad Estoup.

Estoup noted that crypto scammers often use deepfake technology to creat fake videos of well-known personalities to execute scams.

An example of such a scam was a deepfake video of FTX’s former CEO in November, where scammers used old interview footage of Sam Bankman-Fried and a voice emulator to direct users to a malicious website promising to “double your cryptocurrency.”

Schwed said that the volatile nature of crypto causes people to panic and take a “better safe than sorry” approach, which can lead to them getting suckered into deepfake scams. He noted:

“If a video of CZ is released claiming withdrawals will be halted within the hour, are you going to immediately withdraw your funds, or spend hours trying to figure out if the message is real?”

However, Estoup believes that while deepfake technology is advancing at a rapid rate, it’s not yet “indistinguishable from reality.”

How to spot a deepfake: Watch the eyes

Schwed suggests one useful way to quickly spot a deepfake is to watch when the subject blinks their eyes. If it looks unnatural, there’s a good chance it’s a deepfake.

This is due to the fact that deepfakes are generated using image files sourced on the internet, where the subject will usually have their eyes open, explains Schwed. Thus, in a deepfake, the blinking of the subject’s eyes needs to be simulated.

Schwed said the best identifier of course is to ask questions that only the real individual can answer, such as “what restaurant did we meet at for lunch last week?”

Estoup said there is also AI software available that can detect deepfakes and suggests one should look out for big technological improvements in this area.

He also gave some age-old advice: “If it’s too good to be true, it probably is.”

Related: ‘Yikes!’ Elon Musk warns users against latest deepfake crypto scam

Last year, Binance’s chief communications officer, Patrick Hillman, revealed in an August blog post that a sophisticated scam was perpetrated using a deepfake of him.

Hillman noted that the team used previous news interviews and TV appearances over the years to create the deepfake and “fool several highly intelligent crypto members.”

He only became aware of this when he started to receive online messages thanking him for his time talking to project teams about potentially listing their assets on Binance.com.

Earlier this week, blockchain security firm SlowMist noted there were 303 blockchain security incidents in 2022, with 31.6% of them caused by phishing, rug pulls and other scams.