Smooth operators

8Citations
Citations of this article
74Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We develop a generic approach to form smooth versions of basic mathematical operations like multiplication, composition, change of measure, and conditional expectation, among others. Operations which result in functions outside the reproducing kernel Hilbert space (such as the product of two RKHS functions) are approximated via a natural cost function, such that the solution is guaranteed to be in the targeted RKHS. This approximation problem is reduced to a regression problem using an adjoint trick, and solved in a vector-valued RKHS, consisting of continuous, linear, smooth operators which map from an input, real-valued RKHS to the desired target RKHS. Important constraints, such as an almost everywhere positive density, can be enforced or approximated naturally in this framework, using convex constraints on the operators. Finally, smooth operators can be composed to accomplish more complex machine learning tasks, such as the sum rule and kernelized approximate Bayesian inference, where state-of-the-art convergence rates are obtained. Copyright 2013 by the author(s).

Cite

CITATION STYLE

APA

Grünewälder, S., Gretton, A., & Shawe-Taylor, J. (2013). Smooth operators. In 30th International Conference on Machine Learning, ICML 2013 (pp. 2221–2229). International Machine Learning Society (IMLS). https://doi.org/10.12968/s1467-5560(23)60145-9

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free