TY - JOUR
T1 - Information Lattice Learning
AU - Yu, Haizi
AU - Evans, James A.
AU - Varshney, Lav R.
N1 - This work was funded in part by the IBM-Illinois Center for Cognitive Computing Systems Research (C3SR), a research collaboration as part of the IBM AI Horizons Network; the Andrew W. Mellon Foundation on \u201CAlgorithms, Models, and Formalisms\u201D; and the National Science Foundation [#1829366].
PY - 2023
Y1 - 2023
N2 - We propose Information Lattice Learning (ILL) as a general framework to learn rules of a signal (e.g., an image or a probability distribution). In our definition, a rule is a coarsened signal used to help us gain one interpretable insight about the original signal. To make full sense of what might govern the signal's intrinsic structure, we seek multiple disentangled rules arranged in a hierarchy, called a lattice. Compared to representation/rule-learning models optimized for a specific task (e.g., classification), ILL focuses on explainability: it is designed to mimic human experiential learning and discover rules akin to those humans can distill and comprehend. This paper details the math and algorithms of ILL, and illustrates how it addresses the fundamental question "what makes X an X"by creating rule-based explanations designed to help humans understand. Our focus is on explaining X rather than (re)generating it. We present applications in knowledge discovery, using ILL to distill music theory from scores and chemical laws from molecules and further revealing connections between them. We show ILL's efficacy and interpretability on benchmarks and assessments, as well as a demonstration of ILL-enhanced classifiers achieving human-level digit recognition using only one or a few MNIST training examples (1-10 per class).
AB - We propose Information Lattice Learning (ILL) as a general framework to learn rules of a signal (e.g., an image or a probability distribution). In our definition, a rule is a coarsened signal used to help us gain one interpretable insight about the original signal. To make full sense of what might govern the signal's intrinsic structure, we seek multiple disentangled rules arranged in a hierarchy, called a lattice. Compared to representation/rule-learning models optimized for a specific task (e.g., classification), ILL focuses on explainability: it is designed to mimic human experiential learning and discover rules akin to those humans can distill and comprehend. This paper details the math and algorithms of ILL, and illustrates how it addresses the fundamental question "what makes X an X"by creating rule-based explanations designed to help humans understand. Our focus is on explaining X rather than (re)generating it. We present applications in knowledge discovery, using ILL to distill music theory from scores and chemical laws from molecules and further revealing connections between them. We show ILL's efficacy and interpretability on benchmarks and assessments, as well as a demonstration of ILL-enhanced classifiers achieving human-level digit recognition using only one or a few MNIST training examples (1-10 per class).
UR - http://www.scopus.com/inward/record.url?scp=85166240557&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85166240557&partnerID=8YFLogxK
U2 - 10.1613/jair.1.14277
DO - 10.1613/jair.1.14277
M3 - Article
AN - SCOPUS:85166240557
SN - 1076-9757
VL - 77
SP - 971
EP - 1019
JO - Journal of Artificial Intelligence Research
JF - Journal of Artificial Intelligence Research
ER -