<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.0/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" article-type="research-article" xml:lang="en">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">IR</journal-id>
<journal-title-group>
<journal-title>Information Research</journal-title>
</journal-title-group>
<issn pub-type="epub">1368-1613</issn>
<publisher>
<publisher-name>University of Bor&#x00E5;s</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">ir30iConf47209</article-id>
<article-id pub-id-type="doi">10.47989/ir30iConf47209</article-id>
<article-categories>
<subj-group xml:lang="en">
<subject>Research article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Uncovering strategies for identifying deepfakes</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Neo</surname><given-names>Celene</given-names></name>
<xref ref-type="aff" rid="aff0001"/></contrib>
<contrib contrib-type="author"><name><surname>Goh</surname><given-names>Dion Hoe-Lian</given-names></name>
<xref ref-type="aff" rid="aff0002"/></contrib>
<contrib contrib-type="author"><name><surname>Rachel</surname><given-names>Chun Wan Ying</given-names></name>
<xref ref-type="aff" rid="aff0003"/></contrib>
<contrib contrib-type="author"><name><surname>Lee</surname><given-names>Chei Sian</given-names></name>
<xref ref-type="aff" rid="aff0004"/></contrib>
<aff id="aff0001"><bold>Celene Neo</bold> is a final-year Communication Studies student at Nanyang Technological University&#x2019;s Wee Kim Wee School of Communication and Information in Singapore. She can be contacted at <email xlink:href="cneo018@e.ntu.edu.sg">cneo018@e.ntu.edu.sg</email>.</aff>
<aff id="aff0002"><bold>Dion H. Goh</bold> is a Professor at the Wee Kim Wee School of Communication and Information, Nanyang Technological University, Singapore. His research interests include social media practices and perceptions, game-based techniques for shaping user perceptions and motivating behaviour, and online information sharing and seeking. He can be contacted at <email xlink:href="ashlgoh@ntu.edu.sg">ashlgoh@ntu.edu.sg</email>.</aff>
<aff id="aff0003"><bold>Rachel Chun</bold> is a Research Associate at the Wee Kim Wee School of Communication and Information, Nanyang Technological University, Singapore. She is dedicated to extending care and respect to the community through her research and professional endeavours. She can be contacted at <email xlink:href="rachel.chunwy@ntu.edu.sg">rachel.chunwy@ntu.edu.sg</email>.</aff>
<aff id="aff0004"><bold>Chei Sian Lee</bold> is a Professor at the Wee Kim Wee School of Communication and Information, Nanyang Technological University in Singapore. Her research focuses on generative AI and digital nudging, enhancing engagement, decision-making, and learning with an ethical, information- oriented approach. She can be contacted at <email xlink:href="leecs@ntu.edu.sg">leecs@ntu.edu.sg</email>.</aff>
</contrib-group>
<pub-date pub-type="epub"><day>06</day><month>05</month><year>2025</year></pub-date>
<pub-date pub-type="collection"><year>2025</year></pub-date>
<volume>30</volume>
<issue>i</issue>
<fpage>752</fpage>
<lpage>760</lpage>
<permissions>
<copyright-year>2025</copyright-year>
<copyright-holder>&#x00A9; 2025 The Author(s).</copyright-holder>
<license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by-nc/4.0/">
<license-p>This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International License (<ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by-nc/4.0/">http://creativecommons.org/licenses/by-nc/4.0/</ext-link>), permitting all non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<abstract xml:lang="en">
<title>Abstract</title>
<p><bold>Introduction.</bold> The proliferation of generative artificial intelligence tools capable of producing high-quality videos that can masquerade as genuine content has raised concerns about online misinformation. This study investigates human ability to identify deepfake videos, with a focus on identification performance and the strategies employed.</p>
<p><bold>Method.</bold> Data was collected through an online survey. Participants were young adults aged 21 to 35. They were shown four videos and asked to identify them as real or deepfake, followed by questions about the identification strategies used.</p>
<p><bold>Results.</bold> Our results revealed the diverse range of strategies utilised. Predominant strategies centre around assessing the authenticity of traits pertaining to the video&#x2019;s subject as opposed to peripheral details. Furthermore, we uncovered preferences for intuition and strategies that relate to individual decision-making over consulting other individuals or online materials.</p>
<p><bold>Conclusion.</bold> Our results help enhance understanding of how people identify deepfake videos, adding to existing knowledge. These findings also inform initiatives aimed at educating the public about spotting deepfakes. </p>
</abstract>
</article-meta>
</front>
<body>
<sec id="sec1">
<title>Introduction</title>
<p>While generative artificial intelligence (AI) has opened a wealth of creative possibilities, its ability to produce convincing misinformation has made it harder to discern artificial content from real content (<xref rid="R15" ref-type="bibr">Park, 2024</xref>). In recent times, misinformation has taken the form of deepfakes &#x2013; artificially generated videos that manipulate individuals to appear to say or do things they never did (<xref rid="R17" ref-type="bibr">Somers, 2020</xref>). Deepfakes have gained notoriety in the public sphere primarily for their use in generating pornographic content, and more recently, to erode trust by creating high-quality, fabricated videos that falsely depict influential figures making controversial statements.</p>
<p>Existing studies (<xref rid="R7" ref-type="bibr">Heidari et al., 2024</xref>; <xref rid="R9" ref-type="bibr">Jung et al., 2020</xref>; <xref rid="R14" ref-type="bibr">Pan et al., 2020</xref>) on deepfake detection are primarily focused on those involving deep learning. Few studies investigate human deepfake detection strategies. Among those that do, they are often limited in terms of sample size and generalisability due to the use of qualitative data collection methods (<xref rid="R3" ref-type="bibr">Goh et al., 2022</xref>; <xref rid="R21" ref-type="bibr">Zeng et al., 2023</xref>). Videos used in these studies also lack diversity and consist largely of entertainment or political videos (<xref rid="R4" ref-type="bibr">Goh, 2024</xref>), which may not sufficiently encompass the genres of videos that participants encounter day-to-day.</p>
<p>This study addresses these gaps by determining human ability to identify deepfake videos through an online quantitative survey. The first objective ascertains deepfake and real video identification performance; and the second examines the strategies that people employ to identify deepfake and real videos.</p>
</sec>
<sec id="sec2">
<title>Literature review</title>
<p>Deepfakes pose significant threats to society, particularly through their potential to create fabricated videos for spreading misinformation. The misuse of deepfakes risks undermining public trust and has the potential to deepen social divisions (<xref rid="R20" ref-type="bibr">Westerlund, 2019</xref>). The impact of deepfakes is further amplified when considering that individuals generally perceive video content as more credible than textual information (<xref rid="R19" ref-type="bibr">Sundar et al., 2021</xref>), making the spread of falsehoods through deepfakes particularly damaging.</p>
<p>Given the rise in deepfakes (<xref rid="R18" ref-type="bibr">Sumsub, 2023</xref>), there is a wealth of research on deepfake detection using deep learning models (<xref rid="R2" ref-type="bibr">El-Gayar et al., 2024</xref>; <xref rid="R12" ref-type="bibr">Lee et al., 2023</xref>). These approaches enable the rapid processing of large volumes of video data while providing objective assessments (<xref rid="R7" ref-type="bibr">Heidari et al., 2024</xref>), which may explain the heightened research interest over human detection capabilities.</p>
<p>While there is a growing body of work on human deepfake detection strategies, data collection methods are often limited to exploratory and qualitative modes such as interviews (<xref rid="R4" ref-type="bibr">Goh, 2024</xref>) or diary studies (<xref rid="R21" ref-type="bibr">Zeng et al., 2023</xref>). Despite yielding detailed insights, they are constrained by small sample sizes and limited generalisability. Findings are also often limited to the most frequently used strategies regardless of video type. For instance, in Goh (<xref rid="R4" ref-type="bibr">2024</xref>), strategies for identifying video authenticity are described without differentiation between methods used for detecting deepfakes or real videos. This approach limits the ability to understand which specific strategies are most effective for each type of video.</p>
</sec>
<sec id="sec3">
<title>Methodology</title>
<p>Our study utilized an online survey. Participants were shown four videos &#x2014; two real and two deepfake &#x2014; randomly selected from a pool of ten authentic and ten deepfake videos. These videos were publicly available from the Web and covered a range of topics, including entertainment, politics, education, and sports. Video descriptions can be found in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<p>After viewing each video, participants were asked to: (1) identify whether the video was real or deepfake, (2) report their confidence level, and (3) select the strategies they used to arrive at their decision. The strategies were categorized into three types: visual (e.g., facial features, background, and environmental issues), auditory (e.g., vocal features, sound quality), and knowledge-based (e.g., online tools, knowledge of video subject). These strategies were adapted from prior research (<xref rid="R3" ref-type="bibr">Goh et al., 2022</xref>; <xref rid="R21" ref-type="bibr">Zeng et al., 2023</xref>; <xref rid="R4" ref-type="bibr">Goh, 2024</xref>). Participants could select multiple strategies.</p>
<p>Young adults aged 21 to 35 were recruited as they represent a group that is active online and would likely have experience with deepfake videos (<xref rid="R16" ref-type="bibr">Petrosyan, 2024</xref>). A total of 195 participants were recruited via convenience and snowball sampling.</p>
<table-wrap id="T1">
<label>Table 1.</label>
<caption><p>Description of videos used in the study</p></caption>
<table>
<thead>
<tr>
<th align="left" valign="top"><bold>Video Type</bold></th>
<th align="left" valign="top"><bold>Topic</bold></th>
<th align="left" valign="top"><bold>Description</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="10">Deepfake</td>
<td align="left" valign="top" rowspan="4">Entertainment</td>
<td align="left" valign="top">Mark Zuckerberg says he controls billions of peoples&#x2019; confidential data and thus owns their future.</td>
</tr>
<tr>
<td align="left" valign="top">Kim Kardashian tells how she likes making money by manipulating her fans.</td>
</tr>
<tr>
<td align="left" valign="top">Tom Cruise&#x2019;s daily life.</td>
</tr>
<tr>
<td align="left" valign="top"><italic>The Shining</italic> movie clip.</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="3">Politics</td>
<td align="left" valign="top">Manoj Tiwari criticized an opposing political party and encouraged people to vote for his party.</td>
</tr>
<tr>
<td align="left" valign="top">Jeremy Corbyn supports Boris Johnson as Prime Minister.</td>
</tr>
<tr>
<td align="left" valign="top">A speech for the Apollo 11 mission gone wrong.</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="2">Educational</td>
<td align="left" valign="top">Obama reminds people to be more alert to fake news.</td>
</tr>
<tr>
<td align="left" valign="top">Morgan Freeman asks people: Is seeing believing?</td>
</tr>
<tr>
<td align="left" valign="top">Sports</td>
<td align="left" valign="top">Jose Mourinho comments on soccer.</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="10">Real</td>
<td align="left" valign="top" rowspan="4">Entertainment</td>
<td align="left" valign="top">Mark Zuckerberg says he could ascertain people&#x2019;s online behaviours.</td>
</tr>
<tr>
<td align="left" valign="top">Kim Kardashian claims that she cheated on an exam.</td>
</tr>
<tr>
<td align="left" valign="top">Tom Holland taking a break from social media.</td>
</tr>
<tr>
<td align="left" valign="top">Another <italic>The Shining</italic> movie clip.</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="3">Politics</td>
<td align="left" valign="top">Biden criticized MAGA Republicans.</td>
</tr>
<tr>
<td align="left" valign="top">Trump blamed congressional attackers and told his supporters to calm down.</td>
</tr>
<tr>
<td align="left" valign="top">President Uhuru mourns former Kenyan President Mwai Kibaki.</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="2">Educational</td>
<td align="left" valign="top">Hillary Clinton talks about fake news dangers.</td>
</tr>
<tr>
<td align="left" valign="top">Ellen warns people about fake news.</td>
</tr>
<tr>
<td align="left" valign="top">Sports</td>
<td align="left" valign="top">Jose Mourinho on Sir Alex Ferguson&#x2019;s response after Porto&#x2019;s Champions League win over Manchester United.</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec4">
<title>Results</title>
<sec id="sec4_1">
<title>Participant demographics</title>
<p>Participants comprised 83 males and 112 females aged 21 to 35 years. The majority were from fields such as social sciences, finance, engineering, sciences, and computing. YouTube was the most used video platform among participants, followed by Instagram and TikTok. A majority (86%) watched videos daily, with most coming across dubious content <italic>&#x2018;once in a while&#x2019;</italic> when watching videos.</p>
</sec>
<sec id="sec4_2">
<title>Video identification performance</title>
<p>To address the first objective, we found that the distribution of participants skewed towards a higher number of correct identifications. Of 195 participants, the majority (73%) were able to identify more than half of the videos correctly. Only 1% identified zero videos correctly, 6% correctly identified one video, 21% correctly identified two videos, 34% correctly identified three videos, and 38% correctly identified all four videos (<xref ref-type="fig" rid="F1">Figure 1</xref>).</p>
<p><xref ref-type="fig" rid="F2">Figure 2</xref> shows identification accuracy by authenticity type. Here, 53% of participants were able to correctly identify both deepfake videos, while 65% correctly identified both real videos.</p>
<fig id="F1">
<label>Figure 1.</label>
<caption><p>Frequency of number of videos identified correctly</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c64-fig1.jpg"><alt-text>none</alt-text></graphic>
</fig>
<fig id="F2">
<label>Figure 2.</label>
<caption><p>Frequency of correct identifications by authenticity type</p></caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="images\c64-fig2.jpg"><alt-text>none</alt-text></graphic>
</fig>
</sec>
<sec id="sec4_3">
<title>Strategies used in deepfake video identification</title>
<p><xref ref-type="table" rid="T2">Table 2</xref> shows the five most and least used strategies associated with correct identification of deepfake videos. The most frequently used methods typically focused on examining the characteristics of the subject in the video, while the least used methods focused on the peripheral details of the video. Among the most used methods, physical and behavioural characteristics, and intuition and emotions were used half of the time.</p>
<p>Note that the frequency values presented in the tables reflect the total number of times each strategy was employed across both deepfake videos watched (<xref ref-type="table" rid="T2">Tables 2</xref> and <xref ref-type="table" rid="T3">3</xref>) as well as real videos (<xref ref-type="table" rid="T4">Tables 4</xref> and <xref ref-type="table" rid="T5">5</xref>). Each of the 195 participants viewed two real and two deepfake videos, resulting in 390 video assessments per video type. The percentages for each strategy are calculated using this total number of video assessments as the base.</p>
<p><xref ref-type="table" rid="T3">Table 3</xref> shows the top and bottom five strategies associated with incorrect identification of deepfake videos. Strategies were similar to those used in the correct identification of deepfake videos, except for intelligibility and language, and knowledge of video content which were not present in <xref ref-type="table" rid="T2">Table 2</xref>. </p>
<table-wrap id="T2">
<label>Table 2.</label>
<caption><p>Strategies associated with correct deepfake identification</p></caption>
<table>
<thead>
<tr>
<th align="left" valign="top"><bold>Strategy</bold></th>
<th align="left" valign="top"></th>
<th align="left" valign="top"><bold>Frequency</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="5">Most used</td>
<td align="left" valign="top">Physical and behavioural characteristics <italic>(Analysing individuals&#x2019; expressions, body gestures, and movements in the video.)</italic></td>
<td align="left" valign="top">203 (52.1%)</td>
</tr>
<tr>
<td align="left" valign="top">Intuition and emotions <italic>(Based on own instinct, opinions, or emotions.)</italic></td>
<td align="left" valign="top">202 (51.8%)</td>
</tr>
<tr>
<td align="left" valign="top">Vocal features <italic>(Assessing modulation and naturalness in the speaker&#x2019;s voice.)</italic></td>
<td align="left" valign="top">185 (47.4%)</td>
</tr>
<tr>
<td align="left" valign="top">Facial features <italic>(Analysing facial characteristics, such as skin tone, facial symmetry, and hairstyle.)</italic></td>
<td align="left" valign="top">173 (44.4%)</td>
</tr>
<tr>
<td align="left" valign="top">Knowledge of person <italic>(Prior knowledge of people in video.)</italic></td>
<td align="left" valign="top">124 (31.8%)</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="5">Least used</td>
<td align="left" valign="top">Colour and lighting inconsistencies <italic>(Evaluating the video&#x2019;s lighting conditions.)</italic></td>
<td align="left" valign="top">76 (19.5%)</td>
</tr>
<tr>
<td align="left" valign="top">Production issues <italic>(Editing issues, camera angle/work issues, shakiness, and jitter.)</italic></td>
<td align="left" valign="top">67 (17.2%)</td>
</tr>
<tr>
<td align="left" valign="top">Background sound issues <italic>(Background reverberation/echoes, overall noise, mechanical noises.)</italic></td>
<td align="left" valign="top">66 (16.9%)</td>
</tr>
<tr>
<td align="left" valign="top">Use of multiple sources <italic>(Consulting multiple sources when using online tools or communicating with others.)</italic></td>
<td align="left" valign="top">57 (14.6%)</td>
</tr>
<tr>
<td align="left" valign="top">Communication with others <italic>(Checking with family or friends either offline or online.)</italic></td>
<td align="left" valign="top">53 (13.6%)</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T3">
<label>Table 3.</label>
<caption><p>Strategies associated with incorrect deepfake identification</p></caption>
<table>
<thead>
<tr>
<th align="left" valign="top"><bold>Strategy</bold></th>
<th align="left" valign="top"></th>
<th align="left" valign="top"><bold>Frequency</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="5">Most used</td>
<td align="left" valign="top">Vocal features</td>
<td align="left" valign="top">86 (22.1%)</td>
</tr>
<tr>
<td align="left" valign="top">Intuition and emotions</td>
<td align="left" valign="top">72 (18.5%)</td>
</tr>
<tr>
<td align="left" valign="top">Facial features</td>
<td align="left" valign="top">69 (17.7%)</td>
</tr>
<tr>
<td align="left" valign="top">Physical and behavioural characteristics</td>
<td align="left" valign="top">68 (17.4%)</td>
</tr>
<tr>
<td align="left" valign="top">Intelligibility and language <italic>(Assessing the language used, fluency, pronunciation, and intelligibility of speech.)</italic></td>
<td align="left" valign="top">66 (16.9%)</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="5">Least used</td>
<td align="left" valign="top">Knowledge of video content <italic>(Familiarity with events in the video.)</italic></td>
<td align="left" valign="top">32 (8.2%)</td>
</tr>
<tr>
<td align="left" valign="top">Production issues</td>
<td align="left" valign="top">31 (7.9%)</td>
</tr>
<tr>
<td align="left" valign="top">Colour and lighting inconsistencies</td>
<td align="left" valign="top">30 (7.7%)</td>
</tr>
<tr>
<td align="left" valign="top">Use of multiple sources</td>
<td align="left" valign="top">25 (6.4%)</td>
</tr>
<tr>
<td align="left" valign="top">Communication with others</td>
<td align="left" valign="top">21 (5.4%)</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec4_4">
<title>Strategies used in real video identification</title>
<p><xref ref-type="table" rid="T4">Table 4</xref> shows the five most and least used strategies associated with correct identification of real videos. The strategies aligned with those used to correctly identify deepfakes, with the only differing strategy being intelligibility and language.</p>
<p>In terms of the most used strategies, it was interesting to observe that vocal features and facial features saw higher frequency of use as compared to strategies associated with correct identification of deepfake videos. For example, use of vocal features was the predominant strategy, with a frequency of nearly 70%. Compared to <xref ref-type="table" rid="T2">Table 2</xref>, intuition and emotions fell from the second to fifth most used strategy despite maintaining a similar frequency of use. A similar change was observed for physical and behavioural characteristics.</p>
<table-wrap id="T4">
<label>Table 4.</label>
<caption><p>Strategies associated with correct real video identification</p></caption>
<table>
<thead>
<tr>
<th align="left" valign="top"><bold>Strategy</bold></th>
<th align="left" valign="top"></th>
<th align="left" valign="top"><bold>Frequency</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="5">Most used</td>
<td align="left" valign="top">Vocal features</td>
<td align="left" valign="top">267 (68.5%)</td>
</tr>
<tr>
<td align="left" valign="top">Facial features</td>
<td align="left" valign="top">206 (52.8%)</td>
</tr>
<tr>
<td align="left" valign="top">Physical and behavioural characteristics</td>
<td align="left" valign="top">199 (51.0%)</td>
</tr>
<tr>
<td align="left" valign="top">Intelligibility and language</td>
<td align="left" valign="top">190 (48.7%)</td>
</tr>
<tr>
<td align="left" valign="top">Intuition and emotions</td>
<td align="left" valign="top">190 (48.7%)</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="5">Least used</td>
<td align="left" valign="top">Online tools <italic>(Using search engines like Google or social media platforms like Facebook and YouTube.)</italic></td>
<td align="left" valign="top">94 (24.1%)</td>
</tr>
<tr>
<td align="left" valign="top">Production issues</td>
<td align="left" valign="top">76 (19.5%)</td>
</tr>
<tr>
<td align="left" valign="top">Colour and lighting inconsistencies</td>
<td align="left" valign="top">66 (16.9%)</td>
</tr>
<tr>
<td align="left" valign="top">Use of multiple sources</td>
<td align="left" valign="top">52 (13.3%)</td>
</tr>
<tr>
<td align="left" valign="top">Communication with others</td>
<td align="left" valign="top">42 (10.8%)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="T5">Table 5</xref> shows the five most and least used strategies associated with incorrect identification of real videos. Once again, strategies were similar to those used in the correct identification of real videos, which is a common observation between correct and incorrect identification of deepfakes. Notably, this was the only instance where background and environmental details were among the most used strategies.</p>
<table-wrap id="T5">
<label>Table 5.</label>
<caption><p>Strategies associated with incorrect real video identification</p></caption>
<table>
<thead>
<tr>
<th align="left" valign="top"><bold>Strategy</bold></th>
<th align="left" valign="top"></th>
<th align="left" valign="top"><bold>Frequency</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" rowspan="5">Most used</td>
<td align="left" valign="top">Physical and behavioural characteristics</td>
<td align="left" valign="top">49 (12.6%)</td>
</tr>
<tr>
<td align="left" valign="top">Vocal features</td>
<td align="left" valign="top">47 (12.1%)</td>
</tr>
<tr>
<td align="left" valign="top">Intuition and emotions</td>
<td align="left" valign="top">41 (10.5%)</td>
</tr>
<tr>
<td align="left" valign="top">Background and environmental details <italic>(Scene settings, unusual backgrounds, issues with watermarks, logos, or subtitles.)</italic></td>
<td align="left" valign="top">33 (8.5%)</td>
</tr>
<tr>
<td align="left" valign="top">Facial features</td>
<td align="left" valign="top">31 (7.9%)</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="5">Least used</td>
<td align="left" valign="top">Online tools</td>
<td align="left" valign="top">20 (5.1%)</td>
</tr>
<tr>
<td align="left" valign="top">Knowledge of video content</td>
<td align="left" valign="top">19 (4.9%)</td>
</tr>
<tr>
<td align="left" valign="top">Colour and lighting inconsistencies</td>
<td align="left" valign="top">17 (4.4%)</td>
</tr>
<tr>
<td align="left" valign="top">Communication with others</td>
<td align="left" valign="top">17 (4.4%)</td>
</tr>
<tr>
<td align="left" valign="top">Use of multiple sources</td>
<td align="left" valign="top">12 (3.1%)</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="sec5">
<title>Discussion</title>
<p>Overall, participants demonstrated an ability to differentiate between deepfake and authentic videos. When examining performance based on authenticity type, participants exhibited stronger proficiency in detecting real videos over deepfakes.</p>
<p>Across all videos watched, participants used a variety of strategies. This highlights the difficulty in human detection as there is no singular strategy that individuals can rely on to accurately distinguish deepfakes from real videos (<xref rid="R4" ref-type="bibr">Goh, 2024</xref>; <xref rid="R6" ref-type="bibr">Groh, 2020</xref>). This finding thus highlights the importance of media literacy and the use of multiple methods to ascertain the authenticity of videos.</p>
<p>Participants largely focused on the subjects in the video, as seen by facial features, and physical and behavioural characteristics emerging as frequently used strategies. In contrast, strategies that considered general video and audio attributes such as production quality were less frequently utilised. This may stem from the understanding that the subject is often the main focus of deepfake manipulation (<xref rid="R19" ref-type="bibr">Sundar et al., 2021</xref>), prompting participants to ignore other details (<xref rid="R5" ref-type="bibr">Groh et al., 2021</xref>). The danger however is that as the quality of deepfake manipulation increases, people relying primarily on such strategies may fail to spot falsified content.</p>
<p>The strategies used for identifying deepfake videos differed greatly from those used for image and text identification. A study on deepfake images found that participants often relied on peripheral features, such as accessories or clothing texture, to identify manipulated images (<xref rid="R1" ref-type="bibr">Bray et al., 2023</xref>). Similarly, strategies for evaluating the credibility of blog posts emphasised the importance of arrangement and alignment features (<xref rid="R8" ref-type="bibr">Jo et al., 2019</xref>). In contrast, deepfake video detection requires the analysis of dynamic components, such as facial expressions and body movement, which adds a layer of complexity compared to static media types. This distinction illustrates the need for a more nuanced approach to video detection, focusing on dynamic strategies that are not as relevant in text or image analysis.</p>
<p>Vocal features were the most used strategy in the identification of real videos. This suggests that participants perceive the human voice as a convincing feature associated with real videos. The complexity of the human voice and the difficulty of deepfakes to replicate these complexities makes vocal features a critical differentiator (<xref rid="R11" ref-type="bibr">Kulangareth et al., 2024</xref>), aiding participants in distinguishing between real and fabricated content. However as with the concern about visual features, deepfake technology is increasingly able to accurately clone voices (<xref rid="R13" ref-type="bibr">Mai et al., 2023</xref>). Reliance on vocal features only would again pose misidentification dangers.</p>
<p>Notably, regardless of video type and the accuracy of identification, the two least used strategies were employing multiple sources for verification and communication with others. Preference for individual decision-making may be explained by confidence in individual ability to identify deepfakes (K&#x00F6;bis et al, 2021). However, as pointed out by Goh (<xref rid="R4" ref-type="bibr">2024</xref>), the use of more cognitively demanding strategies such as referencing multiple sources increases the likelihood of correct identifications. This is especially important due to the rapid advances in deepfake generation technology.</p>
</sec>
<sec id="sec6">
<title>Conclusion</title>
<p>This study reveals the strategies used in deepfake video identification, addressing a gap in the current literature. The findings underscore the complexity of detecting deepfakes, as evidenced by the array of identification methods utilized.</p>
<p>Our study offers theoretical contributions to the field of human deepfake detection and more generally, information credibility assessment, by revealing the strategies young adults use. In contrast to existing research which often reported on most used strategies (<xref rid="R3" ref-type="bibr">Goh et al., 2022</xref>; <xref rid="R21" ref-type="bibr">Zeng et al., 2023</xref>; <xref rid="R4" ref-type="bibr">Goh, 2024</xref>), our larger participant sample allowed for a detailed analysis of both most and least used ones. The exploration of lesser-preferred strategies offers a more comprehensive understanding of identification approaches and their potential shortcomings. Our findings provide a nuanced, multi-dimensional perspective that enhances existing literature on deepfake detection.</p>
<p>In terms of practical implications, our findings can inform educators in the development of media literacy curricula aimed at enhancing digital wellness and safety. Additionally, authorities and online platforms can use these insights to craft targeted strategies for combating deepfakes, including public service announcements, improving resilience against misinformation.</p>
<p>Despite the insights uncovered, several limitations should be acknowledged. First, the strategies recorded in the survey may not reflect the full suite of those employed by individuals in real world scenarios. Participants were informed that they had the option to rewatch videos multiple times, potentially leading to heightened scrutiny than typical online encounters, inflating the accuracy rates. Future research should aim to capture more naturalistic viewing behaviours. Second, as deepfakes become increasingly sophisticated, new detection strategies may emerge while existing ones may become less effective. Ongoing research is thus essential in providing insights into these evolving methods, ensuring that detection strategies remain relevant and effective in countering ever advancing deepfake technologies.</p>
</sec>
</body>
<back>
<ack>
<title>Acknowledgements</title>
<p>This research project was supported by Nanyang Technological University under the URECA Undergraduate Research Programme as well as a Ministry of Education (Singapore) Tier 2 grant (T2EP40122-0015).</p>
</ack>
<ref-list>
<title>References</title>
<ref id="R1"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bray</surname><given-names>S. D.</given-names></name><name><surname>Johnson</surname><given-names>S. D.</given-names></name><name><surname>Kleinberg</surname><given-names>B.</given-names></name></person-group><year>2023</year><article-title>Testing human ability to detect &#x2018;deepfake&#x2019; images of human faces</article-title><source>Journal of Cybersecurity</source><volume>9</volume><issue>1</issue><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1093/cybsec/tyad011">https://doi.org/10.1093/cybsec/tyad011</ext-link></element-citation></ref>
<ref id="R2"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>El-Gayar</surname><given-names>M.M.</given-names></name><name><surname>Abouhawwash</surname><given-names>M.</given-names></name><name><surname>Askar</surname><given-names>S.S.</given-names></name><name><surname>Sweidan</surname><given-names>S.</given-names></name></person-group><year>2024</year><article-title>A novel approach for detecting deep fake videos using graph neural network</article-title><source>J Big Data</source><volume>11</volume><fpage>22</fpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1186/s40537-024-00884-y">https://doi.org/10.1186/s40537-024-00884-y</ext-link></element-citation></ref>
<ref id="R3"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Goh</surname><given-names>D.H.</given-names></name><name><surname>Lee</surname><given-names>C.S.</given-names></name><name><surname>Chen</surname><given-names>Z.</given-names></name><name><surname>Kuah</surname><given-names>X.W.</given-names></name><name><surname>Pan</surname><given-names>Y.L.</given-names></name></person-group><year>2022</year><article-title>Understanding users&#x2019; deepfake video verification strategies</article-title><source>Communications in Computer and Information Science</source><fpage>1655</fpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1007/978-3-031-19682-9_4">https://doi.org/10.1007/978-3-031-19682-9_4</ext-link></element-citation></ref>
<ref id="R4"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Goh</surname><given-names>D.</given-names></name></person-group><year>2024</year><article-title>He looks very real: Media, knowledge, and search-based strategies for deepfake identification</article-title><source>Journal of the Association for Information Science and Technology</source><volume>75</volume><issue>6</issue><fpage>643</fpage><lpage>654</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1002/asi.24867">https://doi.org/10.1002/asi.24867</ext-link></element-citation></ref>
<ref id="R5"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Groh</surname><given-names>M.</given-names></name><name><surname>Epstein</surname><given-names>Z.</given-names></name><name><surname>Firestone</surname><given-names>C.</given-names></name><name><surname>Picard</surname><given-names>R.</given-names></name></person-group><year>2021</year><article-title>Deepfake detection by human crowds, machines, and machine-informed crowds</article-title><source>Psychological and Cognitive Sciences</source><volume>119</volume><issue>1</issue><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1073/pnas.2110013119">https://doi.org/10.1073/pnas.2110013119</ext-link></element-citation></ref>
<ref id="R6"><element-citation publication-type="book"><person-group person-group-type="author"><name><surname>Groh</surname><given-names>M.</given-names></name></person-group><year>2020</year><source>Detect DeepFakes: How to counteract misinformation created by AI</source><publisher-name>MIT Media Lab</publisher-name><ext-link ext-link-type="uri" xlink:href="https://www.media.mit.edu/projects/detect-fakes/overview/">https://www.media.mit.edu/projects/detect-fakes/overview/</ext-link></element-citation></ref>
<ref id="R7"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Heidari</surname><given-names>A.</given-names></name><name><surname>Navimipour</surname><given-names>N. J.</given-names></name><name><surname>Dag</surname><given-names>H.</given-names></name><name><surname>Unal</surname><given-names>M.</given-names></name></person-group><year>2024</year><article-title>Deepfake detection using deep learning methods: A systematic and comprehensive review</article-title><source>WIREs Data Mining and Knowledge Discovery</source><volume>14</volume><issue>2</issue><fpage>e1520</fpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1002/widm.1520">https://doi.org/10.1002/widm.1520</ext-link></element-citation></ref>
<ref id="R8"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Jo</surname><given-names>Y.</given-names></name><name><surname>Kim</surname><given-names>M.</given-names></name><name><surname>Han</surname><given-names>K.</given-names></name></person-group><year>2019</year><article-title>How Do Humans Assess the Credibility on Web Blogs: Qualifying and Verifying Human Factors with Machine Learning</article-title><source>Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (CHI &#x2019;19)</source><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1145/3290605.3300904">https://doi.org/10.1145/3290605.3300904</ext-link></element-citation></ref>
<ref id="R9"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jung</surname><given-names>T.</given-names></name><name><surname>Kim</surname><given-names>S.</given-names></name><name><surname>Kim</surname><given-names>K.</given-names></name></person-group><year>2020</year><article-title>DeepVision: Deepfakes Detection Using Human Eye Blinking Pattern</article-title><source>IEEE Access</source><volume>8</volume><fpage>83144</fpage><lpage>83154</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/ACCESS.2020.2988660">https://doi.org/10.1109/ACCESS.2020.2988660</ext-link></element-citation></ref>
<ref id="R10"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>K&#x00F6;bis</surname><given-names>N. C.</given-names></name><name><surname>Dolezalov&#x00E1;</surname><given-names>B.</given-names></name><name><surname>Soraperra</surname><given-names>I.</given-names></name></person-group><year>2021</year><article-title>Fooled twice: People cannot detect deepfakes but think they can</article-title><source>iScience</source><volume>24</volume><issue>11</issue><comment>Article 103364</comment><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.isci.2021.103364">https://doi.org/10.1016/j.isci.2021.103364</ext-link></element-citation></ref>
<ref id="R11"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kulangareth</surname><given-names>N.V.</given-names></name><name><surname>Kaufman</surname><given-names>J.</given-names></name><name><surname>Oreskovic</surname><given-names>J.</given-names></name><name><surname>Fossat</surname><given-names>Y.</given-names></name></person-group><year>2024</year><article-title>Investigation of Deepfake Voice Detection Using Speech Pause Patterns: Algorithm Development and Validation</article-title><source>JMIR Biomed Eng</source><volume>9</volume><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.2196/56245">https://doi.org/10.2196/56245</ext-link></element-citation></ref>
<ref id="R12"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname><given-names>E. G.</given-names></name><name><surname>Lee</surname><given-names>I.</given-names></name><name><surname>Yoo</surname><given-names>S. B.</given-names></name></person-group><year>2023</year><article-title>ClueCatcher: Catching Domain-Wise Independent Clues for Deepfake Detection</article-title><source>Mathematics</source><comment>2023</comment><volume>11</volume><issue>18</issue><fpage>3952</fpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/math11183952">https://doi.org/10.3390/math11183952</ext-link></element-citation></ref>
<ref id="R13"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mai</surname><given-names>K. T.</given-names></name><name><surname>Bray</surname><given-names>S.</given-names></name><name><surname>Davies</surname><given-names>T.</given-names></name><name><surname>Griffin</surname><given-names>L. D.</given-names></name></person-group><year>2023</year><article-title>Warning: Humans cannot reliably detect speech deepfakes</article-title><source>PLoS One</source><volume>18</volume><issue>8</issue><fpage>e0285333</fpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1371/journal.pone.0285333">https://doi.org/10.1371/journal.pone.0285333</ext-link></element-citation></ref>
<ref id="R14"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Pan</surname><given-names>D.</given-names></name><name><surname>Sun</surname><given-names>L.</given-names></name><name><surname>Wang</surname><given-names>R.</given-names></name><name><surname>Zhang</surname><given-names>X.</given-names></name><name><surname>Sinnott</surname><given-names>R.O.</given-names></name></person-group><year>2020</year><article-title>Deepfake Detection through Deep Learning</article-title><source>2020 IEEE/ACM International Conference on Big Data Computing, Applications and Technologies (BDCAT),134 - 143</source><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/BDCAT50828.2020.00001">https://doi.org/10.1109/BDCAT50828.2020.00001</ext-link></element-citation></ref>
<ref id="R15"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname><given-names>H. J.</given-names></name></person-group><year>2024</year><article-title>The rise of generative artificial intelligence and the threat of fake news and disinformation online: Perspectives from sexual medicine</article-title><source>Investig Clin Urol</source><volume>65</volume><issue>3</issue><fpage>199</fpage><lpage>201</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.4111/icu.20240015">https://doi.org/10.4111/icu.20240015</ext-link></element-citation></ref>
<ref id="R16"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Petrosyan</surname><given-names>A.</given-names></name></person-group><year>2024</year><article-title>Age distribution of internet users worldwide 2024</article-title><source>Statista</source><ext-link ext-link-type="uri" xlink:href="https://www.statista.com/statistics/272365/age-distribution-of-internet-users-worldwide/">https://www.statista.com/statistics/272365/age-distribution-of-internet-users-worldwide/</ext-link></element-citation></ref>
<ref id="R17"><element-citation publication-type="other"><person-group person-group-type="author"><name><surname>Somers</surname><given-names>M.</given-names></name></person-group><year>2020</year><article-title>Deepfakes, explained</article-title><source>MIT Sloan School of Management</source><ext-link ext-link-type="uri" xlink:href="https://mitsloan.mit.edu/ideas-made-to-matter/deepfakes-explained">https://mitsloan.mit.edu/ideas-made-to-matter/deepfakes-explained</ext-link></element-citation></ref>
<ref id="R18"><element-citation publication-type="book"><person-group person-group-type="author"><collab>Sumsub</collab></person-group><year>2023</year><source>Sumsub Research: Global Deepfake Incidents Surge Tenfold from 2022 to 2023</source><publisher-name>Sumsub</publisher-name><ext-link ext-link-type="uri" xlink:href="https://sumsub.com/newsroom/sumsub-research-global-deepfake-incidents-surge- tenfold-from-2022-to-2023/">https://sumsub.com/newsroom/sumsub-research-global-deepfake-incidents-surge- tenfold-from-2022-to-2023/</ext-link></element-citation></ref>
<ref id="R19"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sundar</surname><given-names>S. S.</given-names></name><name><surname>Molina</surname><given-names>M. D.</given-names></name><name><surname>Cho</surname><given-names>E.</given-names></name></person-group><year>2021</year><article-title>Seeing Is Believing: Is Video Modality More Powerful in Spreading Fake News via Online Messaging Apps</article-title><source>Journal of Computer-Mediated Communication</source><volume>26</volume><issue>6</issue><fpage>301</fpage><lpage>319</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1093/jcmc/zmab010">https://doi.org/10.1093/jcmc/zmab010</ext-link></element-citation></ref>
<ref id="R20"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Westerlund</surname><given-names>M.</given-names></name></person-group><year>2019</year><article-title>The Emergence of Deepfake Technology: A Review</article-title><source>Technology Innovation Management Review</source><volume>9</volume><issue>11</issue><fpage>40</fpage><lpage>53</lpage><ext-link ext-link-type="uri" xlink:href="http://doi.org/10.22215/timreview/1282">http://doi.org/10.22215/timreview/1282</ext-link></element-citation></ref>
<ref id="R21"><element-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zeng</surname><given-names>R.</given-names></name><name><surname>Song</surname><given-names>S.</given-names></name><name><surname>Guo</surname><given-names>Z.</given-names></name><name><surname>Goh</surname><given-names>D.H.</given-names></name><name><surname>Lee</surname><given-names>C.S.</given-names></name></person-group><year>2023</year><article-title>Real or Fake: Eliciting Deepfake Identification Strategies Through a Diary Study</article-title><source>Proceedings of the Association for Information Science and Technology</source><volume>60</volume><issue>1</issue><fpage>1206</fpage><lpage>1208</lpage><ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1002/pra2.993">https://doi.org/10.1002/pra2.993</ext-link></element-citation></ref>
</ref-list>
</back>
</article>