ABSTRACT
|
---|
We give the first algorithm that efficiently learns halfspaces (under distributional assumptions) in the notoriously difficult agnostic framework of Kearns, Schapire, and Sellie. In this model, a learner is given arbitrarily labeled examples from a fixed distribution and must output a hypothesis competitive with the optimal halfspace hypothesis. Our algorithm constructs a hypothesis whose error rate on future examples is within an additive \eps of the optimal halfspace in time poly(n) for any constant \eps>0 under the uniform distribution over {0,1}^n or the unit sphere in R^n, as well as under any log-concave distribution over R^n. It also agnostically learns Boolean disjunctions in time 2^{\tilde{O}(\sqrt{n})} with respect to *any* distribution. The new algorithm, essentially L_1 polynomial regression, is a noise-tolerant arbitrary-distribution generalization of the ``low-degree'' Fourier algorithm of Linial, Mansour, & Nisan. Our Fourier-type algorithm over the unit sphere makes use of approximation properties of various classes of orthogonal polynomials. Time permitting, I will discuss negative results for other halfspace-based classification tasks. Joint work with A. Kalai, Y. Mansour, and R. Servedio. Return to DMTCS home page. |