TY - JOUR
T1 - Fundamental Limits and Tradeoffs in Invariant Representation Learning
AU - Zhao, Han
AU - Dan, Chen
AU - Aragam, Bryon
AU - Jaakkola, Tommi S.
AU - Gordon, Geoffrey J.
AU - Ravikumar, Pradeep
N1 - HZ would like to thank Alexander Kozachinskiy for the discussion in proving Theorem 4.1. HZ and GG would like to acknowledge support from the DARPA XAI project, contract #FA87501720152 and a Nvidia GPU grant. HZ also thanks the support from a Facebook research award. CD and PR would like to acknowledge the support of NSF via IIS-1955532, IIS-1909816, OAC-1934584, and ONR via N000141812861. CD would like to thank Liu Leqi for many helpful discussions in the early stage of this project.
PY - 2022/11/1
Y1 - 2022/11/1
N2 - A wide range of machine learning applications such as privacy-preserving learning, algorithmic fairness, and domain adaptation/generalization among others, involve learning invariant representations of the data that aim to achieve two competing goals: (a) maximize information or accuracy with respect to a target response, and (b) maximize invariance or independence with respect to a set of protected features (e.g. for fairness, privacy, etc). Despite their wide applicability, theoretical understanding of the optimal tradeoffs — with respect to accuracy, and invariance — achievable by invariant representations is still severely lacking. In this paper, we provide an information theoretic analysis of such tradeoffs under both classification and regression settings. More precisely, we provide a geometric characterization of the accuracy and invariance achievable by any representation of the data; we term this feasible region the information plane. We provide an inner bound for this feasible region for the classification case, and an exact characterization for the regression case, which allows us to either bound or exactly characterize the Pareto optimal frontier between accuracy and invariance. Although our contributions are mainly theoretical, a key practical application of our results is in certifying the potential sub-optimality of any given representation learning algorithm for either classification or regression tasks. Our results shed new light on the fundamental interplay between accuracy and invariance, and may be useful in guiding the design of future representation learning algorithms.
AB - A wide range of machine learning applications such as privacy-preserving learning, algorithmic fairness, and domain adaptation/generalization among others, involve learning invariant representations of the data that aim to achieve two competing goals: (a) maximize information or accuracy with respect to a target response, and (b) maximize invariance or independence with respect to a set of protected features (e.g. for fairness, privacy, etc). Despite their wide applicability, theoretical understanding of the optimal tradeoffs — with respect to accuracy, and invariance — achievable by invariant representations is still severely lacking. In this paper, we provide an information theoretic analysis of such tradeoffs under both classification and regression settings. More precisely, we provide a geometric characterization of the accuracy and invariance achievable by any representation of the data; we term this feasible region the information plane. We provide an inner bound for this feasible region for the classification case, and an exact characterization for the regression case, which allows us to either bound or exactly characterize the Pareto optimal frontier between accuracy and invariance. Although our contributions are mainly theoretical, a key practical application of our results is in certifying the potential sub-optimality of any given representation learning algorithm for either classification or regression tasks. Our results shed new light on the fundamental interplay between accuracy and invariance, and may be useful in guiding the design of future representation learning algorithms.
KW - Invariant representation learning
KW - domain adaptation
KW - fairness
KW - information theory
KW - privacy-preservation
UR - http://www.scopus.com/inward/record.url?scp=85148073636&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85148073636&partnerID=8YFLogxK
M3 - Article
AN - SCOPUS:85148073636
SN - 1532-4435
VL - 23
JO - Journal of Machine Learning Research
JF - Journal of Machine Learning Research
M1 - 340
ER -