TY - GEN
T1 - Weathering Ongoing Uncertainty
T2 - 2024 IEEE International Conference on Robotics and Automation, ICRA 2024
AU - Puthumanaillam, Gokul
AU - Liu, Xiangyu
AU - Mehr, Negar
AU - Ornik, Melkior
N1 - This work was supported by U.S. Army ERDC under cooperative agreement W9132T2220004, AFOSR grant FA9550-23-1-0131, ONR grant N00014-23-1-2505, NASA grant 80NSSC21K1030, and NSF CNS-2218759
PY - 2024
Y1 - 2024
N2 - Optimal decision-making presents a significant challenge for autonomous systems operating in uncertain, stochastic and time-varying environments. Environmental variability over time can significantly impact the system's optimal decision making strategy for mission completion. To model such environments, our work combines the previous notion of Time-Varying Markov Decision Processes (TVMDP) with partial observability and introduces Time-Varying Partially Observable Markov Decision Processes (TV-POMDP). We propose a twopronged approach to accurately estimate and plan within the TV-POMDP: 1) Memory Prioritized State Estimation (MPSE), which leverages weighted memory to provide more accurate time-varying transition estimates; and 2) an MPSE-integrated planning strategy that optimizes long-term rewards while accounting for temporal constraint. We validate the proposed framework and algorithms using simulations and hardware, with robots exploring a partially observable, time-varying environments. Our results demonstrate superior performance over standard methods, highlighting the framework's effectiveness in stochastic, uncertain, time-varying domains.
AB - Optimal decision-making presents a significant challenge for autonomous systems operating in uncertain, stochastic and time-varying environments. Environmental variability over time can significantly impact the system's optimal decision making strategy for mission completion. To model such environments, our work combines the previous notion of Time-Varying Markov Decision Processes (TVMDP) with partial observability and introduces Time-Varying Partially Observable Markov Decision Processes (TV-POMDP). We propose a twopronged approach to accurately estimate and plan within the TV-POMDP: 1) Memory Prioritized State Estimation (MPSE), which leverages weighted memory to provide more accurate time-varying transition estimates; and 2) an MPSE-integrated planning strategy that optimizes long-term rewards while accounting for temporal constraint. We validate the proposed framework and algorithms using simulations and hardware, with robots exploring a partially observable, time-varying environments. Our results demonstrate superior performance over standard methods, highlighting the framework's effectiveness in stochastic, uncertain, time-varying domains.
UR - http://www.scopus.com/inward/record.url?scp=85202436442&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85202436442&partnerID=8YFLogxK
U2 - 10.1109/ICRA57147.2024.10610954
DO - 10.1109/ICRA57147.2024.10610954
M3 - Conference contribution
AN - SCOPUS:85202436442
T3 - Proceedings - IEEE International Conference on Robotics and Automation
SP - 4612
EP - 4618
BT - 2024 IEEE International Conference on Robotics and Automation, ICRA 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 13 May 2024 through 17 May 2024
ER -