Feature variance regularization: A simpleway to improve the generalizability of neural networks

2Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

To improve the generalization ability of neural networks, we propose a novel regularization method that regularizes the empirical risk using a penalty on the empirical variance of the features. Intuitively, our approach introduces confusion into feature extraction and prevents the models from learning features that may relate to specific training samples. According to our theoretical analysis, our method encourages models to generate closer feature distributions for the training set and unobservable true data and minimize the expected risk as well, which allows the model to adapt to new samples better. We provide a thorough empirical justification of our approach, and achieves a greater improvement than other regularization methods. The experimental results show the effectiveness of our method on multiple visual tasks, including classification (CIFAR100, ImageNet, fine-grained datasets) and semantic segmentation (Cityscapes).

Cite

CITATION STYLE

APA

Huang, R., Sun, H., Liu, J., Tian, L., Wang, L., Shan, Y., & Wang, Y. (2020). Feature variance regularization: A simpleway to improve the generalizability of neural networks. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 41–90). AAAI press. https://doi.org/10.1609/aaai.v34i04.5840

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free