This work aims to improve the performance of active learning techniques for one-class classification (OCC) via dimensionality reduction (DR) and pre-filtering of the unlabelled input data. In practice, the input data of OCC problems is high-dimensional and often contains significant redundancy of negative examples. Thus, DR is typically an important pre-processing step to address the high-dimensionality challenge. However, the redundancy has not been previously addressed. In this work, we propose a framework to exploit the detected DR basis functions of the instance space in order to filter-out most of the redundant data. Instances are removed or maintained using an adaptive threshold-ing operator depending on their distance to the identified DR basis functions. This reduction in the dimensionality, redundancy and size of the instance space results in significant reduction of the computational complexity of active learning for OCC process. For the preserved instances, their distance to the identified DR basis functions is also used in order to select more efficiently the initial training batch as well as additional instances at each iteration of the active training algorithm. This was done by ensuring that the labelled data always contains nearly uniform representation along the different DR basis functions of the instance space. Experimental results show that applying the DR and pre-filtering steps results in better performance of the active learning for OCC.
CITATION STYLE
Ghazel, M., & Japkowicz, N. (2017). Improving active learning for one-class classification using dimensionality reduction. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 10233 LNAI, pp. 39–44). Springer Verlag. https://doi.org/10.1007/978-3-319-57351-9_4
Mendeley helps you to discover research relevant for your work.