TY - GEN
T1 - On The Fairness of Multitask Representation Learning
AU - Li, Yingcong
AU - Oymak, Samet
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - In the context of multitask learning (MTL), representation learning is often accomplished through a feature-extractor φ that is shared across all tasks. This way, intuitively, the statistical cost of learning φ is collaboratively split across all tasks which enables sample efficiency. In this work, we consider a novel fairness scenario where T tasks can be split into majority and minority groups of sizes Tmaj and Tmin respectively: The group assignments are unknown during MTL and Tmin/Tmaj ratio corresponds to the imbalance level of the problem. We further assume that these groups admit r0, r1-dimensional linear representations which are orthogonal to each other, thus, they would not benefit each other during MTL. Our main finding is that misspecification disproportionately hurts the minority tasks and over-parameterization is key to ensuring fairness of MTL representations. Specifically, we prove that, when we fit a R = r0 dimensional misspecified representation, MTL model achieves small task-averaged risk however it has vanishing explanatory power on minority tasks. Conversely, when we fit a R = r0 + r1 dimensional well-specified representation, MTL model achieves small risks on both majority and minority tasks which are on par with the oracle baseline of training each group individually with the hindsight knowledge of assignments. Finally, we provide experimental results which are consistent with our theoretical findings.
AB - In the context of multitask learning (MTL), representation learning is often accomplished through a feature-extractor φ that is shared across all tasks. This way, intuitively, the statistical cost of learning φ is collaboratively split across all tasks which enables sample efficiency. In this work, we consider a novel fairness scenario where T tasks can be split into majority and minority groups of sizes Tmaj and Tmin respectively: The group assignments are unknown during MTL and Tmin/Tmaj ratio corresponds to the imbalance level of the problem. We further assume that these groups admit r0, r1-dimensional linear representations which are orthogonal to each other, thus, they would not benefit each other during MTL. Our main finding is that misspecification disproportionately hurts the minority tasks and over-parameterization is key to ensuring fairness of MTL representations. Specifically, we prove that, when we fit a R = r0 dimensional misspecified representation, MTL model achieves small task-averaged risk however it has vanishing explanatory power on minority tasks. Conversely, when we fit a R = r0 + r1 dimensional well-specified representation, MTL model achieves small risks on both majority and minority tasks which are on par with the oracle baseline of training each group individually with the hindsight knowledge of assignments. Finally, we provide experimental results which are consistent with our theoretical findings.
KW - fairness
KW - imbalanced data
KW - multitask learning
KW - representation learning
KW - upper/lower bounds
UR - https://www.scopus.com/pages/publications/85177568152
UR - https://www.scopus.com/pages/publications/85177568152#tab=citedBy
U2 - 10.1109/ICASSP49357.2023.10095627
DO - 10.1109/ICASSP49357.2023.10095627
M3 - Conference contribution
AN - SCOPUS:85177568152
T3 - ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
BT - ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing, Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 48th IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2023
Y2 - 4 June 2023 through 10 June 2023
ER -