An Information-Theoretic Framework for Unifying Active Learning Problems

17Citations
Citations of this article
22Readers
Mendeley users who have this article in their library.

Abstract

This paper presents an information-theoretic framework for unifying active learning problems: level set estimation (LSE), Bayesian optimization (BO), and their generalized variant. We first introduce a novel active learning criterion that subsumes an existing LSE algorithm and achieves state-of-the-art performance in LSE problems with a continuous input domain. Then, by exploiting the relationship between LSE and BO, we design a competitive information-theoretic acquisition function for BO that has interesting connections to upper confidence bound and max-value entropy search (MES). The latter connection reveals a drawback of MES which has important implications on not only MES but also on other MES-based acquisition functions. Finally, our unifying information-theoretic framework can be applied to solve a generalized problem of LSE and BO involving multiple level sets in a data-efficient manner. We empirically evaluate the performance of our proposed algorithms using synthetic benchmark functions, a real-world dataset, and in hyperparameter tuning of machine learning models.

Cite

CITATION STYLE

APA

Nguyen, Q. P., Hsiang Low, B. K., & Jaillet, P. (2021). An Information-Theoretic Framework for Unifying Active Learning Problems. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 10B, pp. 9126–9134). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i10.17102

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free