Making sense of deepfakes: Socializing AI and building data literacy on GitHub and YouTube

6Citations
Citations of this article
41Readers
Mendeley users who have this article in their library.
Get full text

Abstract

As a form of synthetic media built on the Internet’s extensive visual datasets with evolving machine learning techniques, deepfakes raise the specter of new types of informational harms and possibilities for image-based abuse. There are calls for three types of defensive response: regulation, technical controls, and improved digital or media literacy. Each is problematic by itself. This article asks what kind of literacy can address deepfake harms, proposing an artificial intelligence (AI) and data literacy framework to explore the potential for social learning with deepfakes and identify sites and methods for intervening in their cultures of production. The article applies contextual qualitative content analysis to explore the most popular GitHub repositories and YouTube accounts teaching “how to deepfake.” The analysis shows that these sites contribute to socializing AI and establishing cultures of social learning, offering potential sites of intervention and pointing to new methods for addressing AI and data harms.

Cite

CITATION STYLE

APA

McCosker, A. (2022). Making sense of deepfakes: Socializing AI and building data literacy on GitHub and YouTube. New Media and Society. https://doi.org/10.1177/14614448221093943

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free