Deepfakes & AI

How Many Businesses Are Affected by Deepfake Fraud?

So-called artificial intelligence is a contentious topic for various reasons, including handling of copyright, power usage, privacy concerns and chatbots like ChatGPT sometimes returning incorrect answers to queries. However, there's one thing that critics and evangelists can agree on: AI increasingly permeates many layers of digital life, from culture to business to politics. As is common with new technologies, an increase of illicit usage of AI in the form of deepfakes is deeply connected to its rise in popularity and, more importantly, accessibility. Deepfakes are AI-generated videos or audio formats that impersonate, for example, celebrities or politicians to spread mis- or disinformation or defraud consumers. Businesses around the world can already feel the impact of this type of artificial content.

A survey by identity verification provider Regula conducted among more than 1,000 experts in fraud detection or prevention from countries like the United States, the United Kingdom, France and Germany shows that a sizable chunk of their companies were targeted by one of three methods of advanced identity fraud. 46 percent of respondents experienced cases of synthetic identity fraud, where a combination of real and fake identity components, like a fake social security number and a real name, address and birth date, was used.

37 percent reported seeing voice deepfakes being used, a recent high-profile example of which from the realm of politics was an artificial President Biden robocalling prospective voters in January to dissuade them from voting in the primaries. Video deepfakes are, as of now, less common, with only 29 percent of respondents having already experienced such fraud attempts. With generative AI companies now focusing on moving pictures, with tools like OpenAI's Sora, this problem could become more pronounced in the coming months and years.

This worry is also reflected in the share of experts surveyed who think these three fraud methods will become a genuine threat, with between 80 and 90 percent agreeing with this assessment. "AI-generated fake identities can be difficult for humans to detect, unless they are specially trained to do so", says Ihar Kliashchou, Chief Technology Officer at Regula. "While neural networks may be useful in detecting deepfakes, they should be used in conjunction with other antifraud measures that focus on physical and dynamic parameters."

Description

This chart show experiences/views of experts on advanced identity fraud methods.

Download Chart
Number of internet identity fraud incidents in Canada 2014-2022
Financial loss due to identity fraud in Canada 2022- Q1 2024
Premium statistics
Internal controls to mitigate fraud risk in the U.S. 2016
Premium statistics
Transaction authentication measures to mitigate fraud risk in the U.S. 2016
Premium statistics
Number of registered fraud cases in Hungary 2010-2022
Online identity fraud offenses in the Netherlands 2012-2019

Any more questions?

Get in touch with us quickly and easily.
We are happy to help!

Do you still have questions?

Feel free to contact us anytime using our contact form or visit our FAQ page.

Statista Content & Design

Need infographics, animated videos, presentations, data research or social media charts?

More Information