TY - JOUR
T1 - Equivariance Discovery by Learned Parameter-Sharing
AU - Yeh, Raymond A.
AU - Hu, Yuan Ting
AU - Hasegawa-Johnson, Mark
AU - Schwing, Alexander G.
N1 - We thank NVIDIA for providing GPUs used for this work. This work was supported in part by NSF under Grant #1718221, 2008387, 2045586, 2106825, MRI #1725729, NIFA award 2020-67021-32799 and Cisco Systems Inc. (Gift Award CG 1377144 - thanks for access to Arcetri). RY is supported by a Google PhD Fellowship.
PY - 2022
Y1 - 2022
N2 - Designing equivariance as an inductive bias into deep-nets has been a prominent approach to build effective models, e.g., a convolutional neural network incorporates translation equivariance. However, incorporating these inductive biases requires knowledge about the equivariance properties of the data, which may not be available, e.g., when encountering a new domain. To address this, we study how to discover interpretable equivariances from data. Specifically, we formulate this discovery process as an optimization problem over a model's parameter-sharing schemes. We propose to use the partition distance to empirically quantify the accuracy of the recovered equivariance. Also, we theoretically analyze the method for Gaussian data and provide a bound on the mean squared gap between the studied discovery scheme and the oracle scheme. Empirically, we show that the approach recovers known equivariances, such as permutations and shifts, on sum of numbers and spatially-invariant data.
AB - Designing equivariance as an inductive bias into deep-nets has been a prominent approach to build effective models, e.g., a convolutional neural network incorporates translation equivariance. However, incorporating these inductive biases requires knowledge about the equivariance properties of the data, which may not be available, e.g., when encountering a new domain. To address this, we study how to discover interpretable equivariances from data. Specifically, we formulate this discovery process as an optimization problem over a model's parameter-sharing schemes. We propose to use the partition distance to empirically quantify the accuracy of the recovered equivariance. Also, we theoretically analyze the method for Gaussian data and provide a bound on the mean squared gap between the studied discovery scheme and the oracle scheme. Empirically, we show that the approach recovers known equivariances, such as permutations and shifts, on sum of numbers and spatially-invariant data.
UR - http://www.scopus.com/inward/record.url?scp=85145873821&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85145873821&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85145873821
SN - 2640-3498
VL - 151
SP - 1527
EP - 1545
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
T2 - 25th International Conference on Artificial Intelligence and Statistics, AISTATS 2022
Y2 - 28 March 2022 through 30 March 2022
ER -