Learning fixed-dimension linear thresholds from fragmented data

2Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

We investigate PAC-learning in a situation in which examples (consisting of an input vector and 0/1 label) have some of the components of the input vector concealed from the learner. This is a special case of restricted focus of attention (RFA) learning. Our interest here is in 1-RFA learning, where only a single component of an input vector is given, for each example. We argue that 1-RFA learning merits special consideration within the wider field of RFA learning. It is the most restrictive form of RFA learning (so that positive results apply in general), and it models a type of "data fusion" scenario, where we have sets of observations from a number of separate sensors, but these sensors are uncorrelated sources. Within this setting we study the well-known class of linear threshold functions, the characteristic functions of Euclidean half-spaces. The sample complexity (i.e., sample-size requirement as a function of the parameters) of this learning problem is affected by the input distribution. We show that the sample complexity is always finite, for any given input distribution, but we also exhibit methods for defining "bad" input distributions for which the sample complexity can grow arbitrarily fast. We identify fairly general sufficient conditions for an input distribution to give rise to sample complexity that is polynomial in the PAC parameters ε-1 and δ-1. We give an algorithm whose sample complexity is polynomial in these parameters and tine dimension (number of input components), for input distributions that satisfy our conditions. The run-time is polynomial in ε-1 and δ-1 provided that the dimension is any constant. We show how to adapt the algorithm to handle uniform misclassification noise. © 2001 Elsevier Science.

References Powered by Scopus

Statistical analysis with missing data

14031Citations
N/AReaders
Get full text

Probability Inequalities for Sums of Bounded Random Variables

5570Citations
N/AReaders
Get full text

A theory of the learnable

3705Citations
N/AReaders
Get full text

Cited by Powered by Scopus

A bound on the precision required to estimate a boolean perceptron from its average satisfying assignment

13Citations
N/AReaders
Get full text

Recognizing parameters of the information security unit implemented by the threshold k-valued function

0Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Goldberg, P. W. (2001). Learning fixed-dimension linear thresholds from fragmented data. Information and Computation, 171(1), 98–122. https://doi.org/10.1006/inco.2001.3059

Readers over time

‘09‘10‘11‘12‘13‘14‘15‘17‘1901234

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 10

77%

Professor / Associate Prof. 1

8%

Lecturer / Post doc 1

8%

Researcher 1

8%

Readers' Discipline

Tooltip

Computer Science 11

85%

Business, Management and Accounting 1

8%

Engineering 1

8%

Save time finding and organizing research with Mendeley

Sign up for free
0