Abstract
Among the most effective methods for uncovering high dimensional unstructured data's generating mechanisms are techniques based on disentangling and learning independent causal mechanisms. However, to identify the disentangled model, previous methods need additional observable variables or do not provide identifiability results. In contrast, this work aims to design an identifiable generative model that approximates the underlying mechanisms from observational data using only self-supervision. Specifically, the generative model uses a degenerate mixture prior to learn mechanisms that generate or transform data. We outline sufficient conditions for an identifiable generative model up to three types of transformations that preserve a coarse-grained disentanglement. Moreover, we propose a self-supervised training method based on these identifiability conditions. We validate our approach on MNIST, FashionMNIST, and Sprites datasets, showing that the proposed method identifies disentangled models - by visualization and evaluating the downstream predictive model's accuracy under environment shifts.
Original language | English (US) |
---|---|
Pages (from-to) | 877-903 |
Number of pages | 27 |
Journal | Proceedings of Machine Learning Research |
Volume | 177 |
State | Published - 2022 |
Event | 1st Conference on Causal Learning and Reasoning, CLeaR 2022 - Eureka, United States Duration: Apr 11 2022 → Apr 13 2022 |
Keywords
- Causal mechanisms
- disentanglement
- generative model
- identifiability
ASJC Scopus subject areas
- Artificial Intelligence
- Software
- Control and Systems Engineering
- Statistics and Probability