X-ray inspection is often an essential part of quality control within quality critical manufacturing industries. Within such industries, X-ray image interpretation is resource intensive and typically conducted by humans. An increased level of automatization would be preferable, and recent advances in artificial intelligence (e.g., deep learning) have been proposed as solutions. However, typically, such solutions are overconfident when subjected to new data far from the training data, so-called out-of-distribution (OOD) data; we claim that safe automatic interpretation of industrial X-ray images, as part of quality control of critical products, requires a robust confidence estimation with respect to OOD data. We explored if such a confidence estimation, an OOD detector, can be achieved by explicit modeling of the training data distribution, and the accepted images. For this, we derived an autoencoder model trained unsupervised on a public dataset with X-ray images of metal fusion welds and synthetic data. We explicitly demonstrate the dangers with a conventional supervised learning-based approach and compare it to the OOD detector. We achieve true positive rates of around (Formula presented.) at false positive rates of around (Formula presented.) on samples similar to the training data and correctly detect some example OOD data.
CITATION STYLE
Lindgren, E., & Zach, C. (2022). Industrial X-ray Image Analysis with Deep Neural Networks Robust to Unexpected Input Data. Metals, 12(11). https://doi.org/10.3390/met12111963
Mendeley helps you to discover research relevant for your work.