Scammers in Russia are increasingly using deepfake technologies to steal money. With the help of neural networks, fraudsters create a realistic video image of a person, on whose behalf they send messages to their relatives and friends, the Central Bank press service reported.
Even famous artists fall for the tricks of fraudsters. Recently, scammers used a deepfake of Larisa Dolina to take out a loan for 50 million rubles.
The regulator emphasized that a message from an acquaintance asking for financial assistance can often be a trick by scammers, and also compiled a list of recommendations to avoid falling for it:
- be sure to call the person on whose behalf the money is being requested and verify the information
- if it is not possible to call, ask a personal question in the message that only this acquaintance knows the answer to
- check the video message for sound defects, monotonous speech of the interlocutor, unnatural facial expressions, these are the most noticeable signs of a deepfake
Earlier it became known that fraudsters began to disguise malicious files as applications based on neural networks for changing voices. Such a fake AI application for changing voices among Russians steals their personal data and gives scammers access to a smartphone or other device.
Read materials on the topic:
AI scammers extort money using fabricated voices of SUSU management