TY - GEN
T1 - Achievability results for learning under communication constraints
AU - Raginsky, Maxim
PY - 2009
Y1 - 2009
N2 - The problem of statistical learning is to construct an accurate predictor of a random variable as a function of a correlated random variable on the basis of an i.i.d. training sample from their joint distribution. Allowable predictors are constrained to lie in some specified class, and the goal is to approach asymptotically the performance of the best predictor in the class. We consider two settings in which the learning agent only has access to rate-limited descriptions of the training data, and present information-theoretic bounds on the predictor performance achievable in the presence of these communication constraints. Our proofs do not assume any separation structure between compression and learning and rely on a new class of operational criteria specifically tailored to joint design of encoders and learning algorithms in rate-constrained settings. These operational criteria naturally lead to a learning-theoretic generalization of the rate-distortion function introduced recently by Kramer and Savari in the context of rate-constrained communication of probability distributions.
AB - The problem of statistical learning is to construct an accurate predictor of a random variable as a function of a correlated random variable on the basis of an i.i.d. training sample from their joint distribution. Allowable predictors are constrained to lie in some specified class, and the goal is to approach asymptotically the performance of the best predictor in the class. We consider two settings in which the learning agent only has access to rate-limited descriptions of the training data, and present information-theoretic bounds on the predictor performance achievable in the presence of these communication constraints. Our proofs do not assume any separation structure between compression and learning and rely on a new class of operational criteria specifically tailored to joint design of encoders and learning algorithms in rate-constrained settings. These operational criteria naturally lead to a learning-theoretic generalization of the rate-distortion function introduced recently by Kramer and Savari in the context of rate-constrained communication of probability distributions.
UR - http://www.scopus.com/inward/record.url?scp=70349274208&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=70349274208&partnerID=8YFLogxK
U2 - 10.1109/ITA.2009.5044957
DO - 10.1109/ITA.2009.5044957
M3 - Conference contribution
AN - SCOPUS:70349274208
SN - 9781424439904
T3 - Information Theory and Applications Workshop, ITA 2009
SP - 272
EP - 279
BT - Information Theory and Applications Workshop, ITA 2009
T2 - Information Theory and Applications Workshop, ITA 2009
Y2 - 8 February 2009 through 13 February 2009
ER -