Algorithmic Political Bias Can Reduce Political Polarization

0Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Does algorithmic political bias contribute to an entrenchment and polarization of political positions? Franke (Philosophy and Technology, 35, 7, 2022) argues that it may do so because the bias involves classifications of people as liberals, conservatives, etc., and individuals often conform to the ways in which they are classified. I provide a novel example of this phenomenon in human–computer interactions and introduce a social psychological mechanism (what I shall call ‘implied political labeling’) that has been overlooked in this context but should be experimentally explored. Furthermore, while Franke proposes that algorithmic political classifications entrench political identities, I contend that they may often produce the opposite result. They can lead people to change in ways that disconfirm the classifications (thus causing ‘looping effects’). Consequently and counterintuitively, algorithmic political bias can in fact decrease political entrenchment and polarization.

Cite

CITATION STYLE

APA

Peters, U. (2022, September 1). Algorithmic Political Bias Can Reduce Political Polarization. Philosophy and Technology. Springer Science and Business Media B.V. https://doi.org/10.1007/s13347-022-00576-6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free