If regression tasks are sampled from a distribution, then the expected error for a future task can be estimated by the average empirical errors on the data of a finite sample of tasks, uniformly over a class of regularizing or pre-processing transformations. The bound is dimension free, justifies optimization of the pre-processing feature-map and explains the circumstances under which learning-to-learn is preferable to single task learning. © 2009 Springer Science+Business Media, LLC.
CITATION STYLE
Maurer, A. (2009). Transfer bounds for linear feature learning. Machine Learning, 75(3), 327–350. https://doi.org/10.1007/s10994-009-5109-7
Mendeley helps you to discover research relevant for your work.