Most people are unable to tell they are watching a ‘deepfake’ video even when they are informed that the content they are watching has been digitally altered, research suggests.
The term “deepfake” refers to a video where artificial intelligence and deep learning – an algorithmic learning method used to train computers – has been used to make a person appear to say something they have not.
Notable examples of it include a manipulated video of Richard Nixon’s Apollo 11 presidential address and Barack Obama insulting Donald Trump – with some researchers suggesting illicit use of the technology could make it the most dangerous form of crime in the future.
In the first experiment, conducted by researchers from the University of Oxford, Brown University, and the Royal Society, participants watched five unaltered videos followed by four unaltered videos and one deepfake – with viewers asked to detect which one is false.
The researchers used videos of Tom Cruise created by VFX artist Chris Ume, which have seen the American actor performing magic tricks and telling jokes about Mikhail Gorbachev in videos uploaded to TikTok.
The second experiment is the same as the first, except the viewers have a content warning telling them that one of the videos will be a deepfake.
Participants who were issued the warning beforehand identified the deepfake in 20 per cent compared to ten per cent who were not, but even with a direct warning over 78 per cent of people could not distinguish the deepfake from authentic content.
“Individuals are no more likely to notice anything out of the ordinary when exposed to a deepfake video of neutral content”, the researchers wrote in a pre-release of the paper, “compared to a control group who viewed only authentic videos.” The paper is expected to be published, and peer reviewed, in a few months.
No matter the participants’ familiarity with Mr Cruise, gender, level of social media use, or their confidence in being able to detect altered video, they all exhibited the same errors.
The only characteristic which significantly correlates with the ability to detect a deepfake was age, the researchers found, with older participants better able to identify the deepfake.
“The difficulty of manually detecting real from fake videos (i.e., with the naked eye) threatens to lower the information value of video media entirely”, the researchers predict.
“As people internalise deepfakes’ capacity to deceive, they will rationally place less trust in all online videos, including authentic content.”
Should this continue in the future people will have to rely on warning labels and content moderation on social media to ensure that deceptive videos and other misinformation does not become endemic on platforms.
That said, Facebook, Twitter, and other sites routinely rely on regular users flagging content to their moderators – a task which could prove difficult if people are unable to tell misinformation and authentic content apart.
Facebook in particular has been criticised repeatedly in the past for not providing enough support for its content moderators and failing to remove false content. Research at New York University and France’s Université Grenoble Alpes found that from August 2020 to January 2021, articles from known purveyors of misinformation received six times as many likes, shares, and interactions as legitimate news articles.
Facebook contended that such research does not show the full picture, as “engagement [with Pages] should not … be confused with how many people actually see it on Facebook”.
The researchers also raised concerns that “such warnings may be written off as politically motivated or biased”, as demonstrated by the conspiracy theories surrounding the COVID-19 vaccine or Twitter’s labelling of former president Trump’s tweets.
The aforementioned deepfake of President Obama calling then-President Trump a “total and complete dipshit” was believed to be accurate by 15 per cent of people in a study from 2020, despite the content itself being “highly improbable”.
A more general distrust of information online is a possible outcome of both deepfakes and content warnings, the researchers caution, and “policymakers should take [that] into account when assessing the costs and benefits of moderating online content.”