Analysis of Regularization in Deep Learning Models on Testbed Architectures

0Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Deep Learning models have come into significant use in the field of biology and healthcare, genomics, medical imaging, EEGs, and electronic medical records [1–4]. In the training these models can be affected due to overfitting, which is mainly due to the fact that Deep Learning models try to adapt as much as possible to the training data, looking for the decrease of the training error which leads to the increase of the validation error. To avoid this, different techniques have been developed to reduce overfitting, among which are the Lasso and Ridge regularization, weight decay, batch normalization, early stopping, data augmentation and dropout. In this research, the impact of the neural network architecture, the batch size and the value of the dropout on the decrease of overfitting, as well as on the time of execution of the tests, is analyzed. As identified in the tests, the neural network architectures with the highest number of hidden layers are the ones that try to adapt to the training data set, which makes them more prone to overfitting.

Author supplied keywords

Cite

CITATION STYLE

APA

Cajicá, F. A. M., García Henao, J. A., Hernández, C. J. B., & Riveill, M. (2021). Analysis of Regularization in Deep Learning Models on Testbed Architectures. In Communications in Computer and Information Science (Vol. 1327, pp. 178–192). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-68035-0_13

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free