TY - JOUR
T1 - Vision-Based Construction Worker Activity Analysis Informed by Body Posture
AU - Roberts, Dominic
AU - Torres Calderon, Wilfredo
AU - Tang, Shuai
AU - Golparvar-Fard, Mani
N1 - Funding Information:
We would like to acknowledge the financial support of National Science Foundation (NSF) Grant No. 1729209 and are indebted to Professor Jun Yang for graciously granting access to her data set of construction worker activity videos for this research. We also appreciate the support of Yunpeng Wang, Amir Ibrahim, and other annotators for technical feedback and assistance on the annotation tool and helping the authors obtain ground-truth data. Any opinions, findings, and conclusion or recommendations expressed in this material are those of the authors and do not necessarily reflect the view of the NSF, industry partners, or professionals mentioned.
Publisher Copyright:
© 2020 American Society of Civil Engineers.
PY - 2020/7/1
Y1 - 2020/7/1
N2 - Activity analysis of construction resources is generally performed by manually observing construction operations either in person or through recorded videos. It is thus prone to observer fatigue and bias and is of limited scalability and cost-effectiveness. Automating this procedure obviates these issues and can allow project teams to focus on performance improvement. This paper introduces a novel deep learning- and vision-based activity analysis framework that estimates and tracks two-dimensional (2D) worker pose and outputs per-frame worker activity labels given input red-green-blue (RGB) video footage of a construction worker operation. We used 317 annotated videos of bricklaying and plastering operations to train and validate the proposed method. This method obtained 82.6% mean average precision (mAP) for pose estimation and 72.6% multiple-object tracking accuracy (MOTA), and 81.3% multiple-object tracking precision (MOTP) for pose tracking. Cross-validation activity analysis accuracy of 78.5% was also obtained. We show that worker pose contributes to activity analysis results. This highlights the potential for using vision-based ergonomics assessment methods that rely on pose in conjunction with the proposed method for assessing the ergonomic viability of individual activities.
AB - Activity analysis of construction resources is generally performed by manually observing construction operations either in person or through recorded videos. It is thus prone to observer fatigue and bias and is of limited scalability and cost-effectiveness. Automating this procedure obviates these issues and can allow project teams to focus on performance improvement. This paper introduces a novel deep learning- and vision-based activity analysis framework that estimates and tracks two-dimensional (2D) worker pose and outputs per-frame worker activity labels given input red-green-blue (RGB) video footage of a construction worker operation. We used 317 annotated videos of bricklaying and plastering operations to train and validate the proposed method. This method obtained 82.6% mean average precision (mAP) for pose estimation and 72.6% multiple-object tracking accuracy (MOTA), and 81.3% multiple-object tracking precision (MOTP) for pose tracking. Cross-validation activity analysis accuracy of 78.5% was also obtained. We show that worker pose contributes to activity analysis results. This highlights the potential for using vision-based ergonomics assessment methods that rely on pose in conjunction with the proposed method for assessing the ergonomic viability of individual activities.
UR - http://www.scopus.com/inward/record.url?scp=85084174319&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85084174319&partnerID=8YFLogxK
U2 - 10.1061/(ASCE)CP.1943-5487.0000898
DO - 10.1061/(ASCE)CP.1943-5487.0000898
M3 - Article
AN - SCOPUS:85084174319
SN - 0887-3801
VL - 34
JO - Journal of Computing in Civil Engineering
JF - Journal of Computing in Civil Engineering
IS - 4
M1 - 04020017
ER -