Gadget

Don’t believe your ears – voice deepfakes have arrived

Voice deepfakes are the latest challenge to online security, from WhatsApp chats through to YouTube videos.

The word “deepfake” is a combination of “deep learning” and “fake”. It has emerged thansk to artificial intelligence and machine learning being used to create compelling fakes of images, video, or audio content. 

To determine whether some audio piece is a fake or a speech of a real human, there are several characteristics to consider: the timbre, manner and intonation of speech. For example, a voice deepfake will give out an unnatural monotony of speech. Illegible speech and strange noises should also alert one to something being dodgy while listening to an audio message or a call.

Deepfake technology in itself is harmless, but in the hands of scammers it can become a dangerous tool with lots of opportunities for deception, defamation or disinformation. There haven’t yet been any mass cases of scams involving voice alteration, but there have been several high-profile cases involving voice deepfakes.

In 2019, scammers used this technology to shake down a UK-based energy firm. In a telephone conversation, the scammer pretended to be the chief executive of the firm’s German parent company and requested an urgent transfer of €220,000 to the account of a certain supplier company. A year later, in 2020, scammers used deepfakes to steal up to $35,000,000 from a Japanese company.

“Currently the technology for creating high-quality deepfakes is not available for widespread use,” says Dmitry Anikin, senior data scientist at Kaspersky. “However, in the future, it may become freely open, which could lead to a surge in related fraud. Most likely, attackers will try to generate voices in real time – to impersonate someone’s relative and lure out money, for example. 

“Such a scenario is not realistic for now: creating high-quality deepfakes involves a lot of limited resources. However, to make a low-quality audio fake, fewer resources are required, and fraudsters can use this. These signs can be helpful to spot this fraud.”

To protect yourself from deepfakes, Kaspersky experts recommend:

Exit mobile version