TY - JOUR
T1 - Identifying Coarse-grained Independent Causal Mechanisms with Self-supervision
AU - Wang, Xiaoyang
AU - Nahrstedt, Klara
AU - Koyejo, Sanmi
N1 - Koyejo acknowledges partial funding from a Sloan Fellowship and a Strategic Research Initiatives award from the University of Illinois at Urbana-Champaign. This work was also funded in part by NSF 2046795, 1909577, 1934986, 1922758, 1827126, 1835834, 2126246, and NIFA award 2020-67021-32799. The authors are very grateful to Dr. Hongbin Pei for proofreading multiple iterations of this manuscript and the anonymous reviewers for the suggestions on improving the presentation of this work.
PY - 2022
Y1 - 2022
N2 - Among the most effective methods for uncovering high dimensional unstructured data's generating mechanisms are techniques based on disentangling and learning independent causal mechanisms. However, to identify the disentangled model, previous methods need additional observable variables or do not provide identifiability results. In contrast, this work aims to design an identifiable generative model that approximates the underlying mechanisms from observational data using only self-supervision. Specifically, the generative model uses a degenerate mixture prior to learn mechanisms that generate or transform data. We outline sufficient conditions for an identifiable generative model up to three types of transformations that preserve a coarse-grained disentanglement. Moreover, we propose a self-supervised training method based on these identifiability conditions. We validate our approach on MNIST, FashionMNIST, and Sprites datasets, showing that the proposed method identifies disentangled models - by visualization and evaluating the downstream predictive model's accuracy under environment shifts.
AB - Among the most effective methods for uncovering high dimensional unstructured data's generating mechanisms are techniques based on disentangling and learning independent causal mechanisms. However, to identify the disentangled model, previous methods need additional observable variables or do not provide identifiability results. In contrast, this work aims to design an identifiable generative model that approximates the underlying mechanisms from observational data using only self-supervision. Specifically, the generative model uses a degenerate mixture prior to learn mechanisms that generate or transform data. We outline sufficient conditions for an identifiable generative model up to three types of transformations that preserve a coarse-grained disentanglement. Moreover, we propose a self-supervised training method based on these identifiability conditions. We validate our approach on MNIST, FashionMNIST, and Sprites datasets, showing that the proposed method identifies disentangled models - by visualization and evaluating the downstream predictive model's accuracy under environment shifts.
KW - Causal mechanisms
KW - disentanglement
KW - generative model
KW - identifiability
UR - http://www.scopus.com/inward/record.url?scp=85164534858&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85164534858&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85164534858
SN - 2640-3498
VL - 177
SP - 877
EP - 903
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
T2 - 1st Conference on Causal Learning and Reasoning, CLeaR 2022
Y2 - 11 April 2022 through 13 April 2022
ER -