Single shot, semantic bounding box detectors, trained in a supervised manner are popular in computer vision-aided visual inspections. These methods have several key limitations: (1) bounding boxes capture too much background, especially when images experience perspective transformation; (2) insufficient domain-specific data and cost to label; and (3) redundant or incorrect detection results on videos or multi-frame data; where it is a nontrivial task to select the best detection and check for outliers. Recent developments in commercial augmented reality and robotic hardware can be leveraged to support inspection tasks. A common capability of the previous is the ability to obtain image sequences and camera poses. In this work, the authors leverage pose information as “prior” to address the limitations of existing supervised learned, single-shot, semantic detectors for the application of visual inspection. The authors propose an unsupervised semantic segmentation method (USP), based on unsupervised learning for image segmentation inspired by differentiable feature clustering coupled with a novel outlier rejection and stochastic consensus mechanism for mask refinement. USP was experimentally validated for a spalling quantification task using a mixed reality headset (Microsoft HoloLens 2). Also, a sensitivity study was conducted to evaluate the performance of USP under environmental or operational variations.
CITATION STYLE
Midwinter, M., Al-Sabbag, Z. A., & Yeum, C. M. (2023). Unsupervised defect segmentation with pose priors. Computer-Aided Civil and Infrastructure Engineering, 38(17), 2455–2471. https://doi.org/10.1111/mice.13001
Mendeley helps you to discover research relevant for your work.