Many approaches at the forefront of structural health monitoring rely on cutting-edge techniques from the field of machine learning. Recently, much interest has been directed towards the study of so-called adversarial examples; deliberate input perturbations that deceive machine learning models while remaining semantically identical. This article demonstrates that data-driven approaches to structural health monitoring are vulnerable to attacks of this kind. In the perfect information or ‘white-box’ scenario, a transformation is found that maps every example in the Los Alamos National Laboratory three-storey structure dataset to an adversarial example. Also presented is an adversarial threat model specific to structural health monitoring. The threat model is proposed with a view to motivate discussion into ways in which structural health monitoring approaches might be made more robust to the threat of adversarial attack.
CITATION STYLE
Champneys, M. D., Green, A., Morales, J., Silva, M., & Mascarenas, D. (2021). On the vulnerability of data-driven structural health monitoring models to adversarial attack. Structural Health Monitoring, 20(4), 1476–1493. https://doi.org/10.1177/1475921720920233
Mendeley helps you to discover research relevant for your work.