We consider the practical advantage of the Bayesian approach over maximum a posteriori methods in its ability to smoothen the landscape of generalization performance measures in the space of hyperparameters, which is vitally important for determining the optimal hyperparameters. The variational method is used to approximate the intractable distribution. Using the leave-one-out error of support vector regression as an example, we demonstrate a further advantage of this method in the analytical estimation of the leave-one-out error, without doing the cross-validation. Comparing our theory with the simulations on both artificial (the "sine" function) and benchmark (the Boston Housing) data sets, we get a good agreement. © Springer-Verlag Berlin Heidelberg 2005.
CITATION STYLE
Gao, Z., & Wong, K. Y. M. (2005). Smooth performance landscapes of the variational Bayesian approach. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 3697 LNCS, pp. 241–246). https://doi.org/10.1007/11550907_38
Mendeley helps you to discover research relevant for your work.