Hartley Spectral Pooling for Deep Learning

4Citations
Citations of this article
43Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In most convolution neural networks (CNNs), downsampling hidden layers is adopted for increasing computation efficiency and the receptive field size. Such operation is commonly called pooling. Maximization and averaging over sliding windows (max/average pooling), and plain downsampling in the form of strided convolution are popular pooling methods. Since the pooling is a lossy procedure, a motivation of our work is to design a new pooling approach for less lossy in the dimensionality reduction. Inspired by the spectral pooling proposed by Rippel et al. [1], we present the Hartley transform based spectral pooling method. The proposed spectral pooling avoids the use of complex arithmetic for frequency representation, in comparison with Fourier pooling. The new approach preserves more structure features for network’s discriminability than max and average pooling. We empirically show the Hartley pooling gives rise to the convergence of training CNNs on MNIST and CIFAR-10 datasets.

Cite

CITATION STYLE

APA

Zhang, H., & Ma, J. (2020). Hartley Spectral Pooling for Deep Learning. CSIAM Transactions on Applied Mathematics, 1(3), 518–529. https://doi.org/10.4208/csiam-am.2020-0018

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free