TY - GEN
T1 - MOSAIC
T2 - 14th ACM Multimedia Systems Conference, MMSys 2023
AU - Gokarn, Ila
AU - Sabbella, Hemanth
AU - Hu, Yigong
AU - Abdelzaher, Tarek
AU - Misra, Archan
N1 - Publisher Copyright:
© 2023 ACM.
PY - 2023/6/7
Y1 - 2023/6/7
N2 - Sustaining high fidelity and high throughput of perception tasks over vision sensor streams on edge devices remains a formidable challenge, especially given the continuing increase in image sizes (e.g., generated by 4K cameras) and complexity of DNN models. One promising approach involves criticality-Aware processing, where the computation is directed selectively to "critical"portions of individual image frames. We introduce MOSAIC, a novel system for such criticality-Aware concurrent processing of multiple vision sensing streams that provides a multiplicative increase in the achievable throughput with negligible loss in perception fidelity. MOSAIC determines critical regions from images received from multiple vision sensors and spatially bin-packs these regions using a novel multi-scale Mosaic Across Scales (MoS) tiling strategy into a single 'canvas frame', sized such that the edge device can retain sufficiently high processing throughput. Experimental studies using benchmark datasets for two tasks, Automatic License Plate Recognition and Drone-based Pedestrian Detection, shows that MOSAIC, executing on a Jetson TX2 edge device, can provide dramatic gains in the throughput vs. fidelity tradeoff. For instance, for drone-based pedestrian detection, for a batch size of 4, MOSAIC can pack input frames from 6 cameras to achieve (a) 4.75X (475%) higher throughput (23 FPS per camera, cumulatively 138FPS) with ≤ 1% accuracy loss, compared to a First Come First Serve (FCFS) processing paradigm.
AB - Sustaining high fidelity and high throughput of perception tasks over vision sensor streams on edge devices remains a formidable challenge, especially given the continuing increase in image sizes (e.g., generated by 4K cameras) and complexity of DNN models. One promising approach involves criticality-Aware processing, where the computation is directed selectively to "critical"portions of individual image frames. We introduce MOSAIC, a novel system for such criticality-Aware concurrent processing of multiple vision sensing streams that provides a multiplicative increase in the achievable throughput with negligible loss in perception fidelity. MOSAIC determines critical regions from images received from multiple vision sensors and spatially bin-packs these regions using a novel multi-scale Mosaic Across Scales (MoS) tiling strategy into a single 'canvas frame', sized such that the edge device can retain sufficiently high processing throughput. Experimental studies using benchmark datasets for two tasks, Automatic License Plate Recognition and Drone-based Pedestrian Detection, shows that MOSAIC, executing on a Jetson TX2 edge device, can provide dramatic gains in the throughput vs. fidelity tradeoff. For instance, for drone-based pedestrian detection, for a batch size of 4, MOSAIC can pack input frames from 6 cameras to achieve (a) 4.75X (475%) higher throughput (23 FPS per camera, cumulatively 138FPS) with ≤ 1% accuracy loss, compared to a First Come First Serve (FCFS) processing paradigm.
KW - canvas-based processing
KW - edge AI
KW - machine perception
UR - http://www.scopus.com/inward/record.url?scp=85163636874&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85163636874&partnerID=8YFLogxK
U2 - 10.1145/3587819.3590986
DO - 10.1145/3587819.3590986
M3 - Conference contribution
AN - SCOPUS:85163636874
T3 - MMSys 2023 - Proceedings of the 14th ACM Multimedia Systems Conference
SP - 278
EP - 288
BT - MMSys 2023 - Proceedings of the 14th ACM Multimedia Systems Conference
PB - Association for Computing Machinery
Y2 - 7 June 2023 through 10 June 2023
ER -