TY - GEN
T1 - Panoptic segmentation forecasting
AU - Graber, Colin
AU - Tsai, Grace
AU - Firman, Michael
AU - Brostow, Gabriel
AU - Schwing, Alexander
N1 - Funding Information:
5. Conclusions We introduced the novel task ‘panoptic segmentation forecasting.’ It requires to anticipate a per-pixel instance-level segmentation of ‘stuff’ and ‘things’ for an unobserved future frame given as input a sequence of past frames. To solve this task, we developed a model which anticipates trajectory and appearance of ‘things’ and by reprojecting input semantics for ‘stuff.’ We demonstrated that the method outperforms compelling baselines on panoptic, semantic and instance segmentation forecasting. Acknowledgements: This work is supported in part by NSF under Grant #1718221, 2008387, 2045586, MRI #1725729, and NIFA award 2020-67021-32799.
Publisher Copyright:
© 2021 IEEE
PY - 2021
Y1 - 2021
N2 - Our goal is to forecast the near future given a set of recent observations. We think this ability to forecast, i.e., to anticipate, is integral for the success of autonomous agents which need not only passively analyze an observation but also must react to it in real-time. Importantly, accurate forecasting hinges upon the chosen scene decomposition. We think that superior forecasting can be achieved by decomposing a dynamic scene into individual 'things' and background 'stuff'. Background 'stuff' largely moves because of camera motion, while foreground 'things' move because of both camera and individual object motion. Following this decomposition, we introduce panoptic segmentation forecasting. Panoptic segmentation forecasting opens up a middle-ground between existing extremes, which either forecast instance trajectories or predict the appearance of future image frames. To address this task we develop a two-component model: one component learns the dynamics of the background stuff by anticipating odometry, the other one anticipates the dynamics of detected things. We establish a leaderboard for this novel task, and validate a state-of-the-art model that outperforms available baselines.
AB - Our goal is to forecast the near future given a set of recent observations. We think this ability to forecast, i.e., to anticipate, is integral for the success of autonomous agents which need not only passively analyze an observation but also must react to it in real-time. Importantly, accurate forecasting hinges upon the chosen scene decomposition. We think that superior forecasting can be achieved by decomposing a dynamic scene into individual 'things' and background 'stuff'. Background 'stuff' largely moves because of camera motion, while foreground 'things' move because of both camera and individual object motion. Following this decomposition, we introduce panoptic segmentation forecasting. Panoptic segmentation forecasting opens up a middle-ground between existing extremes, which either forecast instance trajectories or predict the appearance of future image frames. To address this task we develop a two-component model: one component learns the dynamics of the background stuff by anticipating odometry, the other one anticipates the dynamics of detected things. We establish a leaderboard for this novel task, and validate a state-of-the-art model that outperforms available baselines.
UR - http://www.scopus.com/inward/record.url?scp=85122368622&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85122368622&partnerID=8YFLogxK
U2 - 10.1109/CVPR46437.2021.01233
DO - 10.1109/CVPR46437.2021.01233
M3 - Conference contribution
AN - SCOPUS:85122368622
T3 - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
SP - 12512
EP - 12521
BT - Proceedings - 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2021
PB - IEEE Computer Society
T2 - 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2021
Y2 - 19 June 2021 through 25 June 2021
ER -