Dungeons & Deepfakes: Using scenario-based role-play to study journalists' behavior towards using AI-based verification tools for video content

要旨

The evolving landscape of manipulated media, including the threat of deepfakes, has made information verification a daunting challenge for journalists. Technologists have developed tools to detect deepfakes, but these tools can sometimes yield inaccurate results, raising concerns about inadvertently disseminating manipulated content as authentic news. This study examines the impact of unreliable deepfake detection tools on information verification. We conducted role-playing exercises with 24 US journalists, immersing them in complex breaking-news scenarios where determining authenticity was challenging. Through these exercises, we explored questions regarding journalists' investigative processes, use of a deepfake detection tool, and decisions on when and what to publish. Our findings reveal that journalists are diligent in verifying information, but sometimes rely too heavily on results from deepfake detection tools. We argue for more cautious release of such tools, accompanied by proper training for users to mitigate the risk of unintentionally propagating manipulated content as real news.

著者
Saniat Sohrawardi
Rochester Institute of Technology, Rochester, New York, United States
Matthew Wright
Rochester Institute of Technology, Rochester, New York, United States
Yijing Kelly Wu
Rochester Institute of Technology, Rochester, New York, United States
Andrea Hickerson
The University of Mississippi, Oxford, Mississippi, United States
論文URL

doi.org/10.1145/3613904.3641973

動画

会議: CHI 2024

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2024.acm.org/)

セッション: Privacy and Deepfake

313C
5 件の発表
2024-05-14 20:00:00
2024-05-14 21:20:00