How Good Is My Test Data? Introducing Safety Analysis for Computer Vision

41Citations
Citations of this article
71Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Good test data is crucial for driving new developments in computer vision (CV), but two questions remain unanswered: which situations should be covered by the test data, and how much testing is enough to reach a conclusion? In this paper we propose a new answer to these questions using a standard procedure devised by the safety community to validate complex systems: the hazard and operability analysis (HAZOP). It is designed to systematically identify possible causes of system failure or performance loss. We introduce a generic CV model that creates the basis for the hazard analysis and—for the first time—apply an extensive HAZOP to the CV domain. The result is a publicly available checklist with more than 900 identified individual hazards. This checklist can be utilized to evaluate existing test datasets by quantifying the covered hazards. We evaluate our approach by first analyzing and annotating the popular stereo vision test datasets Middlebury and KITTI. Second, we demonstrate a clearly negative influence of the hazards in the checklist on the performance of six popular stereo matching algorithms. The presented approach is a useful tool to evaluate and improve test datasets and creates a common basis for future dataset designs.

Cite

CITATION STYLE

APA

Zendel, O., Murschitz, M., Humenberger, M., & Herzner, W. (2017). How Good Is My Test Data? Introducing Safety Analysis for Computer Vision. International Journal of Computer Vision, 125(1–3), 95–109. https://doi.org/10.1007/s11263-017-1020-z

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free