ABSTRACT
|
|---|
|
We give the first algorithm that efficiently learns halfspaces (under
distributional assumptions) in the notoriously difficult agnostic framework of
Kearns, Schapire, and Sellie. In this model, a learner is given arbitrarily
labeled examples from a fixed distribution and must output a hypothesis
competitive with the optimal halfspace hypothesis.
Our algorithm constructs a hypothesis whose error rate on future
examples is within an additive \eps of the optimal halfspace in time
poly(n) for any constant \eps>0 under the uniform distribution over
{0,1}^n or the unit sphere in R^n, as well as under any
log-concave distribution over R^n. It also agnostically learns Boolean
disjunctions in time 2^{\tilde{O}(\sqrt{n})} with respect to
*any* distribution. The new algorithm, essentially L_1 polynomial
regression, is a noise-tolerant arbitrary-distribution generalization
of the ``low-degree'' Fourier algorithm of Linial, Mansour, & Nisan. Our
Fourier-type algorithm over the unit sphere makes use of approximation
properties of various classes of orthogonal polynomials.
Time permitting, I will discuss negative results for other halfspace-based classification tasks.
Joint work with A. Kalai, Y. Mansour, and R. Servedio.
Return to DMTCS home page. |