Beyond the low-degree algorithm: Mixtures of subcubes and their applications

25Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.

Abstract

We introduce the problem of learning mixtures of k subcubes over (0,1)n, which contains many classic learning theory problems as a special case (and is itself a special case of others). We give a surprising nO(log k)-time learning algorithm based on higher-order multilinear moments. It is not possible to learn the parameters because the same distribution can be represented by quite different models. Instead, we develop a framework for reasoning about how multilinear moments can pinpoint essential features of the mixture, like the number of components. We also give applications of our algorithm to learning decision trees with stochastic transitions (which also capture interesting scenarios where the transitions are deterministic but there are latent variables). Using our algorithm for learning mixtures of subcubes, we can approximate the Bayes optimal classifier within additive error ϵ on k-leaf decision trees with at most s stochastic transitions on any root-to-leaf path in nO(s+log k) · poly(1/ϵ) time. In this stochastic setting, the classic nO(log k) · poly(1/ϵ)-time algorithms of Rivest, Blum, and Ehrenfreucht-Haussler for learning decision trees with zero stochastic transitions break down because they are fundamentally Occam algorithms. The low-degree algorithm of Linial-Mansour-Nisan is able to get a constant factor approximation to the optimal error (again within an additive ϵ) and runs in time nO(s+log(k/ϵ)). The quasipolynomial dependence on 1/ϵ is inherent to the low-degree approach because the degree needs to grow as the target accuracy decreases, which is undesirable when ϵ is small. In contrast, as we will show, mixtures of k subcubes are uniquely determined by their 2 logk order moments and hence provide a useful abstraction for simultaneously achieving the polynomial dependence on 1/ϵ of the classic Occam algorithms for decision trees and the flexibility of the low-degree algorithm in being able to accommodate stochastic transitions. Using our multilinear moment techniques, we also give the first improved upper and lower bounds since the work of Feldman-O’Donnell-Servedio for the related but harder problem of learning mixtures of binary product distributions.

References Powered by Scopus

Learning Decision Lists

651Citations
N/AReaders
Get full text

Efficient noise-tolerant learning from statistical queries

489Citations
N/AReaders
Get full text

Constant Depth Circuits, Fourier Transform, and Learnability

430Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Top-down induction of decision trees: Rigorous guarantees and inherent limitations

18Citations
N/AReaders
Get full text

Interaction is necessary for distributed learning with privacy or communication constraints

10Citations
N/AReaders
Get full text

Efficiently learning structured distributions from untrusted batches

7Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Chen, S., & Moitra, A. (2019). Beyond the low-degree algorithm: Mixtures of subcubes and their applications. In Proceedings of the Annual ACM Symposium on Theory of Computing (pp. 869–880). Association for Computing Machinery. https://doi.org/10.1145/3313276.3316375

Readers over time

‘19‘20‘21‘22‘23‘2402468

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 5

50%

Researcher 3

30%

Professor / Associate Prof. 1

10%

Lecturer / Post doc 1

10%

Readers' Discipline

Tooltip

Computer Science 7

64%

Mathematics 2

18%

Physics and Astronomy 1

9%

Social Sciences 1

9%

Save time finding and organizing research with Mendeley

Sign up for free
0