Revisiting data augmentation for rotational invariance in convolutional neural networks

6Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Convolutional Neural Networks (CNN) offer state of the art performance in various computer vision tasks. Many of those tasks require different subtypes of affine invariances (scale, rotational, translational) to image transformations. Convolutional layers are translation equivariant by design, but in their basic form lack invariances. In this work we investigate how best to include rotational invariance in a CNN for image classification. Our experiments show that networks trained with data augmentation alone can classify rotated images nearly as well as in the normal unrotated case; this increase in representational power comes only at the cost of training time. We also compare data augmentation versus two modified CNN models for achieving rotational invariance or equivariance, Spatial Transformer Networks and Group Equivariant CNNs, finding no significant accuracy increase with these specialized methods. In the case of data augmented networks, we also analyze which layers help the network to encode the rotational invariance, which is important for understanding its limitations and how to best retrain a network with data augmentation to achieve invariance to rotation.

Cite

CITATION STYLE

APA

Quiroga, F., Ronchetti, F., Lanzarini, L., & Bariviera, A. F. (2020). Revisiting data augmentation for rotational invariance in convolutional neural networks. In Advances in Intelligent Systems and Computing (Vol. 894, pp. 127–141). Springer Verlag. https://doi.org/10.1007/978-3-030-15413-4_10

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free