Identifying best hyperparameters for deep architectures using random forests

0Citations
Citations of this article
4Readers
Mendeley users who have this article in their library.
Get full text

Abstract

A major problem in deep learning is identifying appropriate hyperparameter configurations for deep architectures. This issue is important because: (1) inappropriate hyperparameter configurations will lead to mediocre performance; (2) little expert experience is available to make an informed decision. Random search is a straightforward choice for this problem; however, expensive time cost for each test has made numerous trails impractical. The main strategy of our solution has been based on data modeling via random forest, which is used as a tool to analyze data characteristics of performance of deep architectures with respect to hyperparameter variants and to explore underlying interactions of hyperparameters. This is a general method suitable for all types of deep architecture. Our approach is tested by using deep belief network: the error rate reduced from 1.2% to 0.89% by merely replacing three hyperparameter values.

Cite

CITATION STYLE

APA

Li, Z. Z., Zhong, Z. Y., & Jin, L. W. (2015). Identifying best hyperparameters for deep architectures using random forests. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8994, pp. 29–42). Springer Verlag. https://doi.org/10.1007/978-3-319-19084-6_4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free