Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy

Research output: Contribution to journalArticlepeer-review

Standard Standard

Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy. / McStay, Andrew.
In: Big Data and Society, Vol. 7, No. 1, 2020.

Research output: Contribution to journalArticlepeer-review

HarvardHarvard

APA

CBE

MLA

VancouverVancouver

McStay A. Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy. Big Data and Society. 2020;7(1). Epub 2020 Mar 27. doi: 10.1177/2053951720904386

Author

RIS

TY - JOUR

T1 - Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy

AU - McStay, Andrew

N1 - This work was supported by the UK’s Arts and Humanities Research Council [grant number AH/M006654/1]

PY - 2020

Y1 - 2020

N2 - By the early 2020s, emotional artificial intelligence (emotional AI) will become increasingly present in everyday objects and practices such as assistants, cars, games, mobile phones, wearables, toys, marketing, insurance, policing, education and border controls. There is also keen interest in using these technologies to regulate and optimize the emotional experiences of spaces, such as workplaces, hospitals, prisons, classrooms, travel infrastructures, restaurants, retail and chain stores. Developers frequently claim that their applications do not identify people. Taking the claim at face value, this paper asks, what are the privacy implications of emotional AI practices that do not identify individuals? To investigate privacy perspectives on soft non-identifying emotional AI, the paper draws upon the following: over 100 interviews with the emotion detection industry, legal community, policy-makers, regulators and NGOs interested in privacy; a workshop with stakeholders to design ethical codes for using data about emotions; a UK survey of 2068 citizens on feelings about emotion capture technologies. It finds a weak consensus among social stakeholders on the need for privacy, this driven by different interests and motivations. Given this weak consensus, it concludes that there exists a limited window of opportunity to societally agree principles of practice regarding privacy and the use of data about emotions.

AB - By the early 2020s, emotional artificial intelligence (emotional AI) will become increasingly present in everyday objects and practices such as assistants, cars, games, mobile phones, wearables, toys, marketing, insurance, policing, education and border controls. There is also keen interest in using these technologies to regulate and optimize the emotional experiences of spaces, such as workplaces, hospitals, prisons, classrooms, travel infrastructures, restaurants, retail and chain stores. Developers frequently claim that their applications do not identify people. Taking the claim at face value, this paper asks, what are the privacy implications of emotional AI practices that do not identify individuals? To investigate privacy perspectives on soft non-identifying emotional AI, the paper draws upon the following: over 100 interviews with the emotion detection industry, legal community, policy-makers, regulators and NGOs interested in privacy; a workshop with stakeholders to design ethical codes for using data about emotions; a UK survey of 2068 citizens on feelings about emotion capture technologies. It finds a weak consensus among social stakeholders on the need for privacy, this driven by different interests and motivations. Given this weak consensus, it concludes that there exists a limited window of opportunity to societally agree principles of practice regarding privacy and the use of data about emotions.

KW - Affective computing

KW - biometrics

KW - consensus

KW - data protection

KW - emotional AI

KW - group privacy

U2 - 10.1177/2053951720904386

DO - 10.1177/2053951720904386

M3 - Article

VL - 7

JO - Big Data and Society

JF - Big Data and Society

SN - 2053-9517

IS - 1

ER -