In this paper, we demonstrate how complex deep learning structures can be understood by humans, if likened to isolated but understandable concepts that use the architecture of Nonnegativity Constrained Autoencoder (NCAE). We show that by constraining most of the weights in the network to be nonnegative using both L1 and L2 nonnegativity penalization, a more understandable structure can result with minute deterioration in classification accuracy. Also, this proposed approach yields a more sparse feature extraction and additional output layer sparsification. The concept is illustrated using MNIST and the NORB datasets.
CITATION STYLE
Ayinde, B. O., Hosseini-Asl, E., & Zurada, J. M. (2016). Visualizing and understanding nonnegativity constrained sparse autoencoder in deep learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9692, pp. 3–14). Springer Verlag. https://doi.org/10.1007/978-3-319-39378-0_1
Mendeley helps you to discover research relevant for your work.