TY - GEN
T1 - Artificial mental phenomena
T2 - 3rd ACM Conference on Fairness, Accountability, and Transparency, FAT* 2020
AU - Liang, Lizhen
AU - Acuna, Daniel E.
N1 - Publisher Copyright:
© 2020 Copyright held by the owner/author(s). Publication rights licensed to the Association for Computing Machinery.
PY - 2020/1/27
Y1 - 2020/1/27
N2 - Detecting biases in artificial intelligence has become difficult because of the impenetrable nature of deep learning. The central difficulty is in relating unobservable phenomena deep inside models with observable, outside quantities that we can measure from inputs and outputs. For example, can we detect gendered perceptions of occupations (e.g., female librarian, male electrician) using questions to and answers from a word embedding-based system? Current techniques for detecting biases are often customized for a task, dataset, or method, affecting their generalization. In this work, we draw from Psychophysics in Experimental Psychology-meant to relate quantities from the real world (i.e., “Physics”) into subjective measures in the mind (i.e., “Psyche”)-to propose an intellectually coherent and generalizable framework to detect biases in AI. Specifically, we adapt the two-alternative forced choice task (2AFC) to estimate potential biases and the strength of those biases in black-box models. We successfully reproduce previously-known biased perceptions in word embeddings and sentiment analysis predictions. We discuss how concepts in experimental psychology can be naturally applied to understanding artificial mental phenomena, and how psychophysics can form a useful methodological foundation to study fairness in AI.
AB - Detecting biases in artificial intelligence has become difficult because of the impenetrable nature of deep learning. The central difficulty is in relating unobservable phenomena deep inside models with observable, outside quantities that we can measure from inputs and outputs. For example, can we detect gendered perceptions of occupations (e.g., female librarian, male electrician) using questions to and answers from a word embedding-based system? Current techniques for detecting biases are often customized for a task, dataset, or method, affecting their generalization. In this work, we draw from Psychophysics in Experimental Psychology-meant to relate quantities from the real world (i.e., “Physics”) into subjective measures in the mind (i.e., “Psyche”)-to propose an intellectually coherent and generalizable framework to detect biases in AI. Specifically, we adapt the two-alternative forced choice task (2AFC) to estimate potential biases and the strength of those biases in black-box models. We successfully reproduce previously-known biased perceptions in word embeddings and sentiment analysis predictions. We discuss how concepts in experimental psychology can be naturally applied to understanding artificial mental phenomena, and how psychophysics can form a useful methodological foundation to study fairness in AI.
KW - Artificial Psychophysics
KW - Biases in Sentiment Analysis
KW - Biases in word embeddings
KW - Two-alternative forced choice task
UR - http://www.scopus.com/inward/record.url?scp=85079690171&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85079690171&partnerID=8YFLogxK
U2 - 10.1145/3351095.3375623
DO - 10.1145/3351095.3375623
M3 - Conference contribution
AN - SCOPUS:85079690171
T3 - FAT* 2020 - Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency
SP - 403
EP - 412
BT - FAT* 2020 - Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency
PB - Association for Computing Machinery, Inc
Y2 - 27 January 2020 through 30 January 2020
ER -