TY - JOUR
T1 - Online Continual Adaptation with Active Self-Training
AU - Zhou, Shiji
AU - Zhao, Han
AU - Zhang, Shanghang
AU - Wang, Lianzhe
AU - Chang, Heng
AU - Wang, Zhi
AU - Zhu, Wenwu
N1 - This work is supported by the National Key Research and Development Program of China No. 2020AAA0106300, National Natural Science Foundation of China No. 62050110 and No. 61872215, and the Shenzhen Science and Technology Program of Grant No. RCYX20200714114523079. HZ would like to thank support from a Facebook research award. The authors also thank Kuaishou for sponsoring the research. We gratefully thank Tianyu Liu, Chenghao Hu, Jiangjing Yan, Yinan Mao for proof reading.
PY - 2022
Y1 - 2022
N2 - Models trained with offline data often suffer from continual distribution shifts and expensive labeling in changing environments. This calls for a new online learning paradigm where the learner can continually adapt to changing environments with limited labels. In this paper, we propose a new online setting - Online Active Continual Adaptation, where the learner aims to continually adapt to changing distributions using both unlabeled samples and active queries of limited labels. To this end, we propose Online Self-Adaptive Mirror Descent (OSAMD), which adopts an online teacher-student structure to enable online self-training from unlabeled data, and a margin-based criterion that decides whether to query the labels to track changing distributions. Theoretically, we show that, in the separable case, OSAMD has an O(T2/3) dynamic regret bound under mild assumptions, which is aligned with the Ω(T2/3) lower bound of online learning algorithms with full labels. In the general case, we show a regret bound of O(T2/3 + α∗T), where α∗ denotes the separability of domains and is usually small. Our theoretical results show that OSAMD can fast adapt to changing environments with active queries. Empirically, we demonstrate that OSAMD achieves favorable regrets under changing environments with limited labels on both simulated and real-world data, which corroborates our theoretical findings.
AB - Models trained with offline data often suffer from continual distribution shifts and expensive labeling in changing environments. This calls for a new online learning paradigm where the learner can continually adapt to changing environments with limited labels. In this paper, we propose a new online setting - Online Active Continual Adaptation, where the learner aims to continually adapt to changing distributions using both unlabeled samples and active queries of limited labels. To this end, we propose Online Self-Adaptive Mirror Descent (OSAMD), which adopts an online teacher-student structure to enable online self-training from unlabeled data, and a margin-based criterion that decides whether to query the labels to track changing distributions. Theoretically, we show that, in the separable case, OSAMD has an O(T2/3) dynamic regret bound under mild assumptions, which is aligned with the Ω(T2/3) lower bound of online learning algorithms with full labels. In the general case, we show a regret bound of O(T2/3 + α∗T), where α∗ denotes the separability of domains and is usually small. Our theoretical results show that OSAMD can fast adapt to changing environments with active queries. Empirically, we demonstrate that OSAMD achieves favorable regrets under changing environments with limited labels on both simulated and real-world data, which corroborates our theoretical findings.
UR - http://www.scopus.com/inward/record.url?scp=85163070337&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85163070337&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85163070337
SN - 2640-3498
VL - 151
SP - 8852
EP - 8883
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
T2 - 25th International Conference on Artificial Intelligence and Statistics, AISTATS 2022
Y2 - 28 March 2022 through 30 March 2022
ER -