TY - JOUR
T1 - Democratizing algorithmic news recommenders
T2 - How to materialize voice in a technologically saturated media ecosystem
AU - Harambam, J.
AU - Helberger, N.
AU - Van Hoboken, J.
PY - 2018/11/28
Y1 - 2018/11/28
N2 - © 2018 The Author(s) Published by the Royal Society. All rights reserved.The deployment of various forms of AI, most notably of machine learning algorithms, radically transforms many domains of social life. In this paper we focus on the news industry, where different algorithms are used to customize news offerings to increasingly specific audience preferences. While this personalization of news enables media organizations to be more receptive to their audience, it can be questioned whether current deployments of algorithmic news recommenders (ANR) live up to their emancipatory promise. Like in various other domains, people have little knowledge of what personal data is used and how such algorithmic curation comes about, let alone that they have any concrete ways to influence these data-driven processes. Instead of going down the intricate avenue of trying to make ANR more transparent, we explore in this article ways to give people more influence over the information news recommendation algorithms provide by thinking about and enabling possibilities to express voice. After differentiating four ideal typical modalities of expressing voice (alternation, awareness, adjustment and obfuscation) which are illustrated with currently existing empirical examples, we present and argue for algorithmic recommender personae as a way for people to take more control over the algorithms that curate people's news provision. This article is part of a theme issue 'Governing artificial intelligence: ethical, legal, and technical opportunities and challenges'.
AB - © 2018 The Author(s) Published by the Royal Society. All rights reserved.The deployment of various forms of AI, most notably of machine learning algorithms, radically transforms many domains of social life. In this paper we focus on the news industry, where different algorithms are used to customize news offerings to increasingly specific audience preferences. While this personalization of news enables media organizations to be more receptive to their audience, it can be questioned whether current deployments of algorithmic news recommenders (ANR) live up to their emancipatory promise. Like in various other domains, people have little knowledge of what personal data is used and how such algorithmic curation comes about, let alone that they have any concrete ways to influence these data-driven processes. Instead of going down the intricate avenue of trying to make ANR more transparent, we explore in this article ways to give people more influence over the information news recommendation algorithms provide by thinking about and enabling possibilities to express voice. After differentiating four ideal typical modalities of expressing voice (alternation, awareness, adjustment and obfuscation) which are illustrated with currently existing empirical examples, we present and argue for algorithmic recommender personae as a way for people to take more control over the algorithms that curate people's news provision. This article is part of a theme issue 'Governing artificial intelligence: ethical, legal, and technical opportunities and challenges'.
U2 - 10.1098/rsta.2018.0088
DO - 10.1098/rsta.2018.0088
M3 - Article
SN - 1364-503X
VL - 376
JO - Philosophical Transactions of the Royal Society A Mathematical, Physical and Engineering Sciences
JF - Philosophical Transactions of the Royal Society A Mathematical, Physical and Engineering Sciences
IS - 2133
M1 - 0088
ER -