TY - GEN
T1 - Semantic-Aware View Prediction for 360-Degree Videos at the 5G Edge
AU - Vats, Shivi
AU - Park, Jounsup
AU - Nahrstedt, Klara
AU - Zink, Michael
AU - Sitaraman, Ramesh
AU - Hellwagner, Hermann
N1 - ACKNOWLEDGMENTS This work was supported by the National Science Foundation under Grant No. CNS-1900875 and No. CNS-1901137 and by the 5G Playground Carinthia (https://5gplayground.at/).
PY - 2022
Y1 - 2022
N2 - In a 5G testbed, we use 360° video streaming to test, measure, and demonstrate the 5G infrastructure, including the capabilities and challenges of edge computing support. Specifically, we use the SEAWARE (Semantic-Aware View Prediction) software system, originally described in [1], at the edge of the 5G network to support a 360° video player (handling tiled videos) by view prediction. Originally, SEAWARE performs semantic analysis of a 360° video on the media server, by extracting, e.g., important objects and events. This video semantic information is encoded in specific data structures and shared with the client in a DASH streaming framework. Making use of these data structures, the client/player can perform view prediction without in-depth, computationally expensive semantic video analysis. In this paper, the SEAWARE system was ported and adapted to run (partially) on the edge where it can be used to predict views and prefetch predicted segments/tiles in high quality in order to have them available close to the client when requested. The paper gives an overview of the 5G testbed, the overall architecture, and the implementation of SEAWARE at the edge server. Since an important goal of this work is to achieve low motion-to-glass latencies, we developed and describe "tile postloading", a technique that allows non-predicted tiles to be fetched in high quality into a segment already available in the player buffer. The performance of 360° tiled video playback on the 5G infrastructure is evaluated and presented. Current limitations of the 5G network in use and some challenges of DASH-based streaming and of edge-assisted viewport prediction under "real-world"constraints are pointed out; further, the performance benefits of tile postloading are disclosed.
AB - In a 5G testbed, we use 360° video streaming to test, measure, and demonstrate the 5G infrastructure, including the capabilities and challenges of edge computing support. Specifically, we use the SEAWARE (Semantic-Aware View Prediction) software system, originally described in [1], at the edge of the 5G network to support a 360° video player (handling tiled videos) by view prediction. Originally, SEAWARE performs semantic analysis of a 360° video on the media server, by extracting, e.g., important objects and events. This video semantic information is encoded in specific data structures and shared with the client in a DASH streaming framework. Making use of these data structures, the client/player can perform view prediction without in-depth, computationally expensive semantic video analysis. In this paper, the SEAWARE system was ported and adapted to run (partially) on the edge where it can be used to predict views and prefetch predicted segments/tiles in high quality in order to have them available close to the client when requested. The paper gives an overview of the 5G testbed, the overall architecture, and the implementation of SEAWARE at the edge server. Since an important goal of this work is to achieve low motion-to-glass latencies, we developed and describe "tile postloading", a technique that allows non-predicted tiles to be fetched in high quality into a segment already available in the player buffer. The performance of 360° tiled video playback on the 5G infrastructure is evaluated and presented. Current limitations of the 5G network in use and some challenges of DASH-based streaming and of edge-assisted viewport prediction under "real-world"constraints are pointed out; further, the performance benefits of tile postloading are disclosed.
KW - 5G networks
KW - Tile-based 360° video streaming
KW - edge computing
KW - tile postloading
KW - viewport prediction
UR - http://www.scopus.com/inward/record.url?scp=85147546352&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85147546352&partnerID=8YFLogxK
U2 - 10.1109/ISM55400.2022.00025
DO - 10.1109/ISM55400.2022.00025
M3 - Conference contribution
AN - SCOPUS:85147546352
T3 - Proceedings - 2022 IEEE International Symposium on Multimedia, ISM 2022
SP - 121
EP - 128
BT - Proceedings - 2022 IEEE International Symposium on Multimedia, ISM 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 24th IEEE International Symposium on Multimedia, ISM 2022
Y2 - 5 December 2022 through 7 December 2022
ER -