TY - GEN
T1 - Towards Intersectionality in Machine Learning
T2 - 5th ACM Conference on Fairness, Accountability, and Transparency, FAccT 2022
AU - Wang, Angelina
AU - Ramaswamy, Vikram V.
AU - Russakovsky, Olga
N1 - Publisher Copyright:
© 2022 Owner/Author.
PY - 2022/6/21
Y1 - 2022/6/21
N2 - Research in machine learning fairness has historically considered a single binary demographic attribute; however, the reality is of course far more complicated. In this work, we grapple with questions that arise along three stages of the machine learning pipeline when incorporating intersectionality as multiple demographic attributes: (1) which demographic attributes to include as dataset labels, (2) how to handle the progressively smaller size of subgroups during model training, and (3) how to move beyond existing evaluation metrics when benchmarking model fairness for more subgroups. For each question, we provide thorough empirical evaluation on tabular datasets derived from the US Census, and present constructive recommendations for the machine learning community. First, we advocate for supplementing domain knowledge with empirical validation when choosing which demographic attribute labels to train on, while always evaluating on the full set of demographic attributes. Second, we warn against using data imbalance techniques without considering their normative implications and suggest an alternative using the structure in the data. Third, we introduce new evaluation metrics which are more appropriate for the intersectional setting. Overall, we provide substantive suggestions on three necessary (albeit not sufficient!) considerations when incorporating intersectionality into machine learning.
AB - Research in machine learning fairness has historically considered a single binary demographic attribute; however, the reality is of course far more complicated. In this work, we grapple with questions that arise along three stages of the machine learning pipeline when incorporating intersectionality as multiple demographic attributes: (1) which demographic attributes to include as dataset labels, (2) how to handle the progressively smaller size of subgroups during model training, and (3) how to move beyond existing evaluation metrics when benchmarking model fairness for more subgroups. For each question, we provide thorough empirical evaluation on tabular datasets derived from the US Census, and present constructive recommendations for the machine learning community. First, we advocate for supplementing domain knowledge with empirical validation when choosing which demographic attribute labels to train on, while always evaluating on the full set of demographic attributes. Second, we warn against using data imbalance techniques without considering their normative implications and suggest an alternative using the structure in the data. Third, we introduce new evaluation metrics which are more appropriate for the intersectional setting. Overall, we provide substantive suggestions on three necessary (albeit not sufficient!) considerations when incorporating intersectionality into machine learning.
UR - http://www.scopus.com/inward/record.url?scp=85133024072&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85133024072&partnerID=8YFLogxK
U2 - 10.1145/3531146.3533101
DO - 10.1145/3531146.3533101
M3 - Conference contribution
AN - SCOPUS:85133024072
T3 - ACM International Conference Proceeding Series
SP - 336
EP - 349
BT - Proceedings of 2022 5th ACM Conference on Fairness, Accountability, and Transparency, FAccT 2022
PB - Association for Computing Machinery
Y2 - 21 June 2022 through 24 June 2022
ER -