TY - GEN
T1 - Improving predictive state representations via gradient descent
AU - Jiang, Nan
AU - Kulesza, Alex
AU - Singh, Satinder
N1 - This work was supported by NSF grant IIS 1319365. Any opinions, findings, conclusions, or recommendations expressed here are those of the authors and do not necessarily reflect the views of the sponsors. We thank Amirreza Shaban and Byron Boots for generously sharing their code and answering questions regarding their method.
PY - 2016
Y1 - 2016
N2 - Predictive state representations (PSRs) model dynamical systems using appropriately chosen predictions about future observations as a representation of the current state. In contrast to the hidden states posited by HMMs or RNNs, PSR states are directly observable in the training data; this gives rise to a moment-matching spectral algorithm for learning PSRs that is computationally efficient and statistically consistent when the model complexity matches that of the true system generating the data. In practice, however, model mismatch is inevitable and while spectral learning remains appealingly fast and simple it may fail to find optimal models. To address this problem, we investigate the use of gradient methods for improving spectrally-learned PSRs. We show that only a small amount of additional gradient optimization can lead to significant performance gains, and moreover that initializing gradient methods with the spectral learning solution yields better models in significantly less time than starting from scratch.
AB - Predictive state representations (PSRs) model dynamical systems using appropriately chosen predictions about future observations as a representation of the current state. In contrast to the hidden states posited by HMMs or RNNs, PSR states are directly observable in the training data; this gives rise to a moment-matching spectral algorithm for learning PSRs that is computationally efficient and statistically consistent when the model complexity matches that of the true system generating the data. In practice, however, model mismatch is inevitable and while spectral learning remains appealingly fast and simple it may fail to find optimal models. To address this problem, we investigate the use of gradient methods for improving spectrally-learned PSRs. We show that only a small amount of additional gradient optimization can lead to significant performance gains, and moreover that initializing gradient methods with the spectral learning solution yields better models in significantly less time than starting from scratch.
UR - https://www.scopus.com/pages/publications/85007197276
UR - https://www.scopus.com/pages/publications/85007197276#tab=citedBy
M3 - Conference contribution
AN - SCOPUS:85007197276
T3 - 30th AAAI Conference on Artificial Intelligence, AAAI 2016
SP - 1709
EP - 1715
BT - 30th AAAI Conference on Artificial Intelligence, AAAI 2016
PB - American Association for Artificial Intelligence (AAAI) Press
T2 - 30th AAAI Conference on Artificial Intelligence, AAAI 2016
Y2 - 12 February 2016 through 17 February 2016
ER -