Carnegie Mellon University
Browse

Polynomial regression under arbitrary product distributions

Download (263.61 kB)
journal contribution
posted on 2006-01-01, 00:00 authored by Eric Blais, Ryan O'Donnell, Karl Wimmer
In recent work, Kalai, Klivans,Mansour, and Servedio [KKMS05] studied a variant of the “Low-Degree (Fourier) Algorithm” for learning under the uniform probability distribution on {0, 1}n. They showed that the L1 polynomial regression algorithm yields agnostic (tolerant to arbitrary noise) learning algorithms with respect to the class of threshold functions — under certain restricted instance distributions, including uniform on {0, 1}n and Gaussian on Rn. In this work we show how all learning results based on the Low-Degree Algorithm can be generalized to give almost identical agnostic guarantees under arbitrary product distributions on instance spaces X1 ×· · ·×Xn . We also extend these results to learning under mixtures of product distributions. The main technical innovation is the use of (Hoeffding) orthogonal decomposition and the extension of the “noise sensitivity method” to arbitrary product spaces. In particular, we give a very simple proof that threshold functions over arbitrary product spaces have δ-noise sensitivity O(√δ), resolving an open problem suggested by Peres [Per04].

History

Date

2006-01-01

Usage metrics

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC