TY - GEN
T1 - Information-theoretic considerations in batch reinforcement learning
AU - Chen, Jinglin
AU - Jiang, Nan
N1 - Publisher Copyright:
© 2019 by the Author(S).
PY - 2019
Y1 - 2019
N2 - Value-function approximation methods that operate in batch mode have foundational importance to reinforcement learning (RL). Finite sample guarantees for these methods often crucially rely on two types of assumptions: (1) mild distribution shift, and (2) representation conditions that arc stronger than realizability. However, the necessity ("why do we need them?") and the naturalness ("when do they hold?") of such assumptions have largely eluded the literature. In this paper, we revisit these assumptions and provide theoretical results towards answering the above questions, and make steps towards a deeper understanding of value-function approximation.
AB - Value-function approximation methods that operate in batch mode have foundational importance to reinforcement learning (RL). Finite sample guarantees for these methods often crucially rely on two types of assumptions: (1) mild distribution shift, and (2) representation conditions that arc stronger than realizability. However, the necessity ("why do we need them?") and the naturalness ("when do they hold?") of such assumptions have largely eluded the literature. In this paper, we revisit these assumptions and provide theoretical results towards answering the above questions, and make steps towards a deeper understanding of value-function approximation.
UR - http://www.scopus.com/inward/record.url?scp=85078047111&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85078047111&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85078047111
T3 - 36th International Conference on Machine Learning, ICML 2019
SP - 1792
EP - 1817
BT - 36th International Conference on Machine Learning, ICML 2019
PB - International Machine Learning Society (IMLS)
T2 - 36th International Conference on Machine Learning, ICML 2019
Y2 - 9 June 2019 through 15 June 2019
ER -