Damage-Detection Approach for Bridges with Multi-Vehicle Loads Using Convolutional Autoencoder
- Authors
- Lee, K[Lee, Kanghyeok]; Jeong, S[Jeong, Seunghoo]; Sim, SH[Sim, Sung-Han]; Shin, D[Shin, Do Hyoung]
- Issue Date
- Mar-2022
- Publisher
- MDPI
- Keywords
- damage detection; convolutional autoencoder; multi-vehicle loads; rigid-frame bridge; reinforced-concrete-slab bridge; deep learning
- Citation
- SENSORS, v.22, no.5
- Indexed
- SCIE
SCOPUS
- Journal Title
- SENSORS
- Volume
- 22
- Number
- 5
- URI
- https://scholarworks.bwise.kr/skku/handle/2021.sw.skku/95970
- DOI
- 10.3390/s22051839
- ISSN
- 1424-8220
- Abstract
- Deep learning has been widely employed in recent studies on bridge-damage detection to improve the performance of damage-detection methods. Unsupervised deep learning can be effectively utilized to increase the applicability of damage-detection approaches. Hence, the authors propose a convolutional-autoencoder (CAE)-based damage-detection approach, which is an unsupervised deep-learning network. However, the CAE-based damage-detection approach demonstrates only satisfactory accuracy for prestressed concrete bridges with a single-vehicle load. Therefore, this study was performed to verify whether the CAE-based damage-detection approach can be applied to bridges with multi-vehicle loads, which is a typical scenario. In this study, rigid-frame and reinforced-concrete-slab bridges were modeled and simulated to obtain the behavior data of bridges. A CAE-based damage-detection approach was tested on both bridges. For both bridges, the results demonstrated satisfactory damage-detection accuracy of over 90% and a false-negative rate of less than 1%. These results prove that the CAE-based approach can be successfully applied to various types of bridges with multi-vehicle loads.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - Engineering > Civil, Architectural Engineering and Landscape Architecture > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.