TY - JOUR
T1 - Automated damage detection for open-air and underwater navigation infrastructure using generative AI-produced training data for deep learning models
AU - Alexander, Quincy G.
AU - Narazaki, Yasutaka
AU - Maxwell, Andrew
AU - Wang, Shengyi
AU - Spencer, Billie F.
N1 - The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: Financial support for this research was provided in part by the US Army Corps of Engineers through a subaward from the University of California, San Diego (Contract/Purchase Order No. W912HZ-17-2-0024).
PY - 2024
Y1 - 2024
N2 - Research has been continually growing toward the development of computer vision-based inspection tools for large-scale civil infrastructure; however, many deep learning techniques require large datasets to properly train models. Collecting field data can be costly and time-consuming, or may not be feasible, which has led to efforts to leverage synthetic data to supplement field data. Recent advances in text-to-image generative artificial intelligence (AI) offer the potential to quickly create realistic synthetic images of damaged infrastructure, including the complexities of the environment found in the field. In this study, the use of text-to-image generation to create a multiclass synthetic training dataset for inland navigation infrastructure is proposed, including damage of underwater structural components. Images of steel and concrete were generated that are representative of inland navigation infrastructure components. The images were labeled for semantic segmentation, and a model was trained using open-to-air and underwater scenes. The model trained using synthetic images was tested against field images, and the performance measured using recall, precision, and intersection over union was found to be comparable to a model trained using only field images. These results demonstrate that text-to-image generative AI tools were shown to be effective for generation of synthetic images with specifically defined conditions, saving time and cost, while providing a similar performance as the use of field-collected images. While intended for damage detection in large-scale civil infrastructure, this concept could be expanded to a number of areas as the generative AI models continue to improve.
AB - Research has been continually growing toward the development of computer vision-based inspection tools for large-scale civil infrastructure; however, many deep learning techniques require large datasets to properly train models. Collecting field data can be costly and time-consuming, or may not be feasible, which has led to efforts to leverage synthetic data to supplement field data. Recent advances in text-to-image generative artificial intelligence (AI) offer the potential to quickly create realistic synthetic images of damaged infrastructure, including the complexities of the environment found in the field. In this study, the use of text-to-image generation to create a multiclass synthetic training dataset for inland navigation infrastructure is proposed, including damage of underwater structural components. Images of steel and concrete were generated that are representative of inland navigation infrastructure components. The images were labeled for semantic segmentation, and a model was trained using open-to-air and underwater scenes. The model trained using synthetic images was tested against field images, and the performance measured using recall, precision, and intersection over union was found to be comparable to a model trained using only field images. These results demonstrate that text-to-image generative AI tools were shown to be effective for generation of synthetic images with specifically defined conditions, saving time and cost, while providing a similar performance as the use of field-collected images. While intended for damage detection in large-scale civil infrastructure, this concept could be expanded to a number of areas as the generative AI models continue to improve.
KW - Synthetic data
KW - automated damage detection
KW - computer vision
KW - generative AI
KW - semantic segmentation
KW - structural health monitoring
KW - text-to-image
UR - http://www.scopus.com/inward/record.url?scp=85213295735&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85213295735&partnerID=8YFLogxK
U2 - 10.1177/14759217241295380
DO - 10.1177/14759217241295380
M3 - Article
AN - SCOPUS:85213295735
SN - 1475-9217
JO - Structural Health Monitoring
JF - Structural Health Monitoring
ER -