Fredholm multiple kernel learning for semi-supervised domain adaptation

11Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

As a fundamental constituent of machine learning, domain adaptation generalizes a learning model from a source domain to a different (but related) target domain. In this paper, we focus on semi-supervised domain adaptation and explicitly extend the applied range of unlabeled target samples into the combination of distribution alignment and adaptive classifier learning. Specifically, our extension formulates the following aspects in a single optimization: 1) learning a cross-domain predictive model by developing the Fredholm integral based kernel prediction framework; 2) reducing the distribution difference between two domains; 3) exploring multiple kernels to induce an optimal learning space. Correspondingly, such an extension is distinguished with allowing for noise resiliency, facilitating knowledge transfer and analyzing diverse data characteristics. It is emphasized that we prove the differentiability of our formulation and present an effective optimization procedure based on the reduced gradient, guaranteeing rapid convergence. Comprehensive empirical studies verify the effectiveness of the proposed method.

Cite

CITATION STYLE

APA

Wang, W., Wang, H., Zhang, C., & Gao, Y. (2017). Fredholm multiple kernel learning for semi-supervised domain adaptation. In 31st AAAI Conference on Artificial Intelligence, AAAI 2017 (pp. 2732–2738). AAAI press. https://doi.org/10.1609/aaai.v31i1.10818

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free