Study shows humans do not easily detect deepfakes

Illustration of the two sets of videos presented to participants. Credit: Computers in Human Behavior (2023). DOI: 10.1016/j.chb.2023.107917

A study of people’s ability to detect “deepfakes” has shown humans perform fairly poorly, even when given hints on how to identify video-based deceit.

Dr. Klaire Somoray and Dr. Dan J Miller from James Cook University have published a new study in Computers in Human Behavior. They state that high-quality deepfake videos, in which a person in an existing image or video is manipulated to have another person’s likeness, can now be generated with ease.

“This has raised concerns about this technology being used for nefarious purposes such as creating political misinformation. For instance, in March 2022, a manipulated video of Ukrainian President Volodymyr Zelensky was circulated, in which Zelensky is depicted appealing for Ukrainian soldiers to surrender,” said Dr. Somoray.

Dr. Somoray and Dr. Miller recruited more than 450 people and showed them 20 videos, 10 of which were real and 10 of which were deepfakes. Participants were then graded on their ability to judge which videos were real, and which were not.

Half of the volunteers were given training on how to spot a deepfake video.

“This includes paying attention to things such as lighting, whether the cheeks and forehead looked too smooth or wrinkly, whether the agedness of the skin was similar to the agedness of the hair and eyes and whether facial hair looked real,” the researchers said.

On average, participants correctly identified approximately 12 out of 20 videos.

“The poorest performers correctly categorized 5 out of 20 videos and the best performers correctly categorized 19 out of 20. Teaching people detection strategies did not impact detection accuracy or detection confidence, nor did time spent per video, or the average number of page clicks on each video,” said Dr. Somoray.

“The findings cast doubt on whether simply providing the public with strategies for detecting deepfakes can meaningfully improve detection.

“Also, worryingly, it appears that individuals may be overly optimistic regarding their abilities to ascertain the authenticity of individual videos,” said Dr. Miller.

More information:
Klaire Somoray et al, Providing detection strategies to improve human detection of deepfakes: An experimental study, Computers in Human Behavior (2023). DOI: 10.1016/j.chb.2023.107917

Provided by
James Cook University


Citation:
Study shows humans do not easily detect deepfakes (2023, September 12)
retrieved 12 September 2023
from https://phys.org/news/2023-09-humans-easily-deepfakes.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.

 

Reference

Denial of responsibility! Samachar Central is an automatic aggregator of Global media. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, and all materials to their authors. For any complaint, please reach us at – [email protected]. We will take necessary action within 24 hours.
Denial of responsibility! Samachar Central is an automatic aggregator of Global media. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, and all materials to their authors. For any complaint, please reach us at – [email protected]. We will take necessary action within 24 hours.
DMCA compliant image

Leave a Comment