Reflecting on Algorithmic Bias With Design Fiction: The MiniCoDe Workshops

3Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

Abstract

In an increasingly complex everyday life, algorithms-often learned from data, i.e., machine learning (ML)-are used to make or assist with operational decisions. However, developers and designers usually are not entirely aware of how to reflect on social justice while designing ML algorithms and applications. Algorithmic social justice-i.e., designing algorithms including fairness, transparency, and accountability-aims at helping expose, counterbalance, and remedy bias and exclusion in future ML-based decision-making applications. How might we entice people to engage in more reflective practices that examine the ethical consequences of ML algorithmic bias in society? We developed and tested a design-fiction-driven methodology to enable multidisciplinary teams to perform intense, workshop-like gatherings to let potential ethical issues emerge and mitigate bias through a series of guided steps. With this contribution, we present an original and innovative use of design fiction as a method to reduce algorithmic bias in co-design activities.

Cite

CITATION STYLE

APA

Turchi, T., Malizia, A., & Borsci, S. (2024). Reflecting on Algorithmic Bias With Design Fiction: The MiniCoDe Workshops. IEEE Intelligent Systems, 39(2), 40–50. https://doi.org/10.1109/MIS.2024.3352977

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free