Uniform-Distribution Attribute Noise Learnability

Similar documents
Uniform-Distribution Attribute Noise Learnability

Lecture 4: LMN Learning (Part 2)

Polynomial time Prediction Strategy with almost Optimal Mistake Probability

3 Finish learning monotone Boolean functions

1 Last time and today

Separating Models of Learning from Correlated and Uncorrelated Data

Yale University Department of Computer Science

Can PAC Learning Algorithms Tolerate. Random Attribute Noise? Sally A. Goldman. Department of Computer Science. Washington University

More Efficient PAC-learning of DNF with Membership Queries. Under the Uniform Distribution

On the Sample Complexity of Noise-Tolerant Learning

CSC 2429 Approaches to the P vs. NP Question and Related Complexity Questions Lecture 2: Switching Lemma, AC 0 Circuit Lower Bounds

TTIC An Introduction to the Theory of Machine Learning. Learning from noisy data, intro to SQ model

Distribution Free Learning with Local Queries

Discriminative Learning can Succeed where Generative Learning Fails

On the Efficiency of Noise-Tolerant PAC Algorithms Derived from Statistical Queries

An Introduction to Statistical Theory of Learning. Nakul Verma Janelia, HHMI

Lecture 29: Computational Learning Theory

On Learning Monotone DNF under Product Distributions

Simple Learning Algorithms for. Decision Trees and Multivariate Polynomials. xed constant bounded product distribution. is depth 3 formulas.

New Results for Random Walk Learning

Being Taught can be Faster than Asking Questions

6.842 Randomness and Computation April 2, Lecture 14

Poly-logarithmic independence fools AC 0 circuits

Learning Functions of Halfspaces Using Prefix Covers

Learning DNF from Random Walks

Lecture 7: Passive Learning

Separating Models of Learning with Faulty Teachers

Boosting and Hard-Core Sets

From Batch to Transductive Online Learning

Learning Juntas. Elchanan Mossel CS and Statistics, U.C. Berkeley Berkeley, CA. Rocco A. Servedio MIT Department of.

Lecture 5: Efficient PAC Learning. 1 Consistent Learning: a Bound on Sample Complexity

Lecture 5: February 16, 2012

Learning DNF Expressions from Fourier Spectrum

Beating the Hold-Out: Bounds for K-fold and Progressive Cross-Validation

Impagliazzo s Hardcore Lemma

Learning and Fourier Analysis

Efficient Noise-Tolerant Learning from Statistical Queries

Lecture 10: Learning DNF, AC 0, Juntas. 1 Learning DNF in Almost Polynomial Time

1 Differential Privacy and Statistical Query Learning

Notes for Lecture 15

Some Selected Topics. Spectral Norm in Learning Theory:

Proclaiming Dictators and Juntas or Testing Boolean Formulae

1 Randomized Computation

Unconditional Lower Bounds for Learning Intersections of Halfspaces

ICML '97 and AAAI '97 Tutorials

On Noise-Tolerant Learning of Sparse Parities and Related Problems

arxiv: v2 [cs.lg] 17 Apr 2013

On the power and the limits of evolvability. Vitaly Feldman Almaden Research Center

A Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997

EQUIVALENCES AND SEPARATIONS BETWEEN QUANTUM AND CLASSICAL LEARNABILITY

COMS 4771 Introduction to Machine Learning. Nakul Verma

On Exact Learning from Random Walk

The sample complexity of agnostic learning with deterministic labels

ON SENSITIVITY OF k-uniform HYPERGRAPH PROPERTIES

Two Sides of the Coin Problem

Computational Learning Theory

Computational Learning Theory. CS534 - Machine Learning

Last time, we described a pseudorandom generator that stretched its truly random input by one. If f is ( 1 2

Learning Combinatorial Functions from Pairwise Comparisons

The information-theoretic value of unlabeled data in semi-supervised learning

Agnostic Learning of Disjunctions on Symmetric Distributions

Introduction to Bayesian Learning. Machine Learning Fall 2018

On Basing Lower-Bounds for Learning on Worst-Case Assumptions

The PAC Learning Framework -II

Quantum algorithms (CO 781/CS 867/QIC 823, Winter 2013) Andrew Childs, University of Waterloo LECTURE 13: Query complexity and the polynomial method

8.1 Polynomial Threshold Functions

Optimal Cryptographic Hardness of Learning Monotone Functions

Testing by Implicit Learning

Learning Sparse Perceptrons

Hierarchical Concept Learning

Tight Bounds on Proper Equivalence Query Learning of DNF

Maximum Margin Algorithms with Boolean Kernels

Computational Learning Theory - Hilary Term : Introduction to the PAC Learning Framework

CS Communication Complexity: Applications and New Directions

Noise-tolerant learning, the parity problem, and the Statistical Query model

Learning symmetric non-monotone submodular functions

On the Sensitivity of Cyclically-Invariant Boolean Functions

PAC Learning. prof. dr Arno Siebes. Algorithmic Data Analysis Group Department of Information and Computing Sciences Universiteit Utrecht

Foundations of Machine Learning and Data Science. Lecturer: Avrim Blum Lecture 9: October 7, 2015

Constructing Hard Functions from Learning Algorithms

Notes for Lecture 25

Pseudorandom Generators

Web-Mining Agents Computational Learning Theory

Lecture 4: Proof of Shannon s theorem and an explicit code

Compute the Fourier transform on the first register to get x {0,1} n x 0.

Lecture Learning infinite hypothesis class via VC-dimension and Rademacher complexity;

PRGs for space-bounded computation: INW, Nisan

: On the P vs. BPP problem. 18/12/16 Lecture 10

Learning and Fourier Analysis

Machine Learning

Learning Intersections and Thresholds of Halfspaces

COS598D Lecture 3 Pseudorandom generators from one-way functions

Lower Bounds for Testing Bipartiteness in Dense Graphs

FORMULATION OF THE LEARNING PROBLEM

Polynomial Certificates for Propositional Classes

Learning Large-Alphabet and Analog Circuits with Value Injection Queries

Computational Learning Theory. Definitions

Online Learning versus Offline Learning*

Trace Reconstruction Revisited

Short Exponent Diffie-Hellman Problems

Transcription:

Uniform-Distribution Attribute Noise Learnability Nader H. Bshouty Dept. Computer Science Technion Haifa 32000, Israel bshouty@cs.technion.ac.il Jeffrey C. Jackson Math. & Comp. Science Dept. Duquesne University Pittsburgh, PA 15282-1754, U.S.A. jackson@mathcs.duq.edu Christino Tamon Dept. Math & Comp. Science Clarkson University Potsdam, NY 13699-5815, U.S.A. tino@clarkson.edu Abstract We study the problem of PAC-learning Boolean functions with random attribute noise under the uniform distribution. First, we define a noisy distance measure for function classes and show that if this measure is small for a class C and an attribute noise distribution D then C is not learnable with respect to the uniform distribution in the presence of noise generated according to D. The noisy distance measure is then characterized in terms of Fourier properties of the function class. We use this characterization to show that the class of all parity functions is not learnable for any but very concentrated noise distributions D. On the other hand, we show that if C is learnable with respect to uniform using a standard Fourier-based learning technique, then C is learnable with time and sample complexity also determined by the noisy distance. In fact, we show that this style algorithm is the best possible for learning in the presence of attribute noise. 1 INTRODUCTION The problem of attribute noise in PAC-learning was studied originally by Shackelford and Volper [8] for the case of k- DNF expressions. Their uniform attribute noise model consists of a Bernoulli process that will either flip or not flip each attribute value with a fixed probability p 2 [0; 1] that is the same for every attribute. While Shackelford and Volper assumed that the learner knows the noise rate p, Goldman and Sloan [5] proved that this assumption is not necessary in order to learn monomials. In addition to uniform attribute noise, Goldman and Sloan also considered a product noise model in which there are n This material is based upon work supported by the National Science Foundation under Grant No. CCR-9800029. noise rates p i, one for each distinct attribute x i, i 2 [n]. They showed that if the product noise rates p i are unknown, then no PAC-learning algorithm exists that can tolerate a noise rate higher than 2, where is the required-accuracy parameter for PAC learning. Their proof uses the method of induced distribution of Kearns and Li [6]. Subsequently, Decatur and Gennaro [2] proved that if the different noise rates are known (or if some upper bound on them is given) then there exist efficient PAC-learning algorithms for simple classes such as monomials and k-dnf. In this paper we consider a very general attribute noise model, but limit the distribution that will be used to generate examples and to evaluate the accuracy of the hypothesis generated by the learning algorithm. Specifically, we focus on the problem of PAC learning with respect to the uniform distribution over examples, with little or no constraint on the distribution used to generate attribute noise in the examples. We give both lower and upper bounds. First, we define a measure of noisy distance for concept classes and show that the sample size required for PAC learning a class over the uniform distribution is inversely proportional to the noisy distance of the class. We also give a characterization of the noisy distance in terms of Fourier properties of the class. As an example of how this characterization can be used, we show that the class of all parity functions is not (even information theoretically) PAC learnable with respect to uniform unless the attribute noise distribution puts nonnegligible weight on one or more noise vectors. This holds even if the noise process is known. On the other hand, we observe as a corollary of a result of Blum, Burch, and Langford [1] that the class of monotone Boolean functions is weakly PAC-learnable even if the noise process if unknown. We then turn to developing positive learnability results. Specifically, we show that any concept class that is PAClearnable with respect to the uniform distribution using an algorithm in the style of Linial, Mansour, and Nisan [7] can be adapted to handle attribute noise, assuming the probability distribution of the noise process is known. However, the noisy distance of a class depends on the noise distribution, so the sample complexity of our algorithm is dependent on the noise process as well as the usual PAC factors. The dependence of the the sample complexity of our algorithm matches, to within polynomial factors, our lower bound for learning with attribute noise. We also give a simple argument showing that if the noise process is unknown then even extremely simple concept classes are not (strongly) learnable.

2 DEFINITIONS AND NOTATION The problem considered in this paper is PAC learning Boolean functions under some fixed distribution over instances when attribute noise is also applied to the instances. To a lesser extent, we also consider classification noise. We now define these concepts more precisely below. For simplicity, our definition suppresses some details of standard definitions (particularly the notion of size of functions) that are not critical to the results in this paper. For a natural number n, we consider classes of Boolean functions f : f0; 1g n! f?1; +1g and distributions over f0; 1g n. The uniform distribution on f0; 1g n is denoted U, i.e., U (x) 2?n, for all x 2 f0; 1g n. The bitwise exclusiveor of two n-bit vectors a; b 2 f0; 1g n is denoted a b. The unit vector e i 2 f0; 1g n has its i-th bit set to one and all other bits set to zero. For a 2 f0; 1g n, the parity function n a is defined as a (x) (?1)P i1 aixi. It is known that any Boolean function f : f0; 1g n! f?1; +1g can be represented as a weighted sum of parity functions (see [7]) f (x) a2f0; 1g n ^f (a) a (x) where ^f (a) E U [f (x) a (x)] is the Fourier coefficient of f at a. This is called the Fourier representation of f and is a direct consequence of the fact that f a j a 2 f0; 1g n g forms an orthonormal basis for all Boolean (or even real-valued) functions over f0; 1g n, i.e., E U [ a (x) b (x)] is one if a b and zero otherwise. The focus of the paper is on a learning model in which the instance distribution is uniform and the noise process is characterized by a pair of parameters (D; R). The noise process can be viewed as drawing a random vector a from the distribution D (representing the attribute noise process) and a random value b from the distribution R (representing classification noise), then returning the exclusive OR of a with the original example vector x and the exclusive OR of the label f (x) with b. So the noise process changes an example (x; f (x)) to an example (x a; f (x) b) (actually, because we consider functions mapping to f?1; +1g, we will assume that R produces values in f?1; +1g and replace the latter with multiplication). We will call this (D; R)-noise and denote the oracle that returns a (D; R)-noisy example for f with respect to the uniform distribution by E D;R (f; U ). Definition 1 Let C be a concept class containing functions f : f0; 1g n! f?1; +1g. Then C is PAC learnable under the uniform distribution with (D; R)-noise if there is an algorithm A such that for any ; 2 (0; 1) and for any target f 2 C, given the inputs ; and access to a noisy example oracle E D;R (f; U ), the algorithm A outputs a hypothesis h such that Pr U [h 6 f ] < with probability at least 1?. The algorithm must make a number of oracle calls (have sample complexity) at most polynomial in n, 1, and 1. The time complexity of A is the number of computation steps taken by A. A PAC algorithm is efficient if its time complexity is also polynomial in n, 1, and 1. If the classification noise process R always returns 0, then (D; R)-noise is simply attribute noise and we refer to it as D-noise. Our lower bounds focus on this type of noise. 3 MODEL TRANSFORMATION Before developing our main results, it is useful to relate the (D; R)-noise model to another model where the example (x; f (x)) is changed to (x; f (x a)b) for a random vector a drawn according to distribution D and b 2 f?1; +1g drawn according to distribution R. Lemma 1 Let U U n be the uniform distribution over n- bit vectors and (D; R) (D n ; R n ) be any distribution over f0; 1g n and f?1; +1g, respectively. Let f : f0; 1g n! f0; 1g be any Boolean function. If 2 U f0; 1g n, A 2 D f0; 1g n and B 2 R f?1; +1g are random variables then the random variables ( A; f ()B) and (; f ( A)B) have identical distributions. Proof Consider the random variables 1 (; A; B) and 2 ( A; A; B). Since is uniformly distributed, 1 and 2 are identically distributed. Define '(x; y; z) (x; f (x y)z). Then ( A; f ()B) '( 2 ) '( 1 ) (; f ( A)B); completing the claim. This lemma is key to our subsequent results, as it allows us to consider the easier noise model of (; f ( A)B) instead of the random attribute noise model when learning is with respect to the uniform distribution. 4 SAMPLE COMPLEITY LOWER BOUND In this section we give a lower bound for PAC-learning with D-noise. Because D-noise is a special case of (D; R)-noise, our lower bounds immediately generalize to this model as well. We start with some intuition for the lower bound. Let C be the class being learned. Let f and g be two functions in the class C and suppose Pr U [f 6 g] >. If for a fixed x and distribution D the expectation E ad [f (x a)] is very close to E ad [g(xa)], then we cannot notice the difference between (x; f (xa 1 )) and (x; g(xa 2 )). Now since the example oracle we consider chooses x according to the uniform distribution, we will look at E x [je a [f (x a)? g(x a)]j]. This, we will show, is a good measure for learnability with noise. We now formalize the above. Definition 2 Let C be a concept class over f0; 1g n and let f; g 2 C. Let D be any distribution over f0; 1g n. Then the noisy distance between f and g under the distribution D is defined as D (f; g) 1 2 E x[je a [f (x a)? g(x a)]j]; where the expectation of x is taken over the uniform distribution over f0; 1g n and the expectation of a is taken with respect to D. For a concept class C let D (C) minf D(f; g) j f; g 2 C with Pr[f 6 g] > g: U The following theorem states an information-theoretic lower bound on the number of samples required by any PAC learning algorithm.

Theorem 2 Let C be a concept class and, for fixed and D, represent D (C) by. Then any PAC learning algorithm for C under a D-distribution noise that, with probability at least 1?, outputs? an -good hypothesis requires a sample 1? complexity of. Proof Consider an algorithm that tries to distinguish whether a sample S fh~x i ; b i i j i 2 [m]g is labeled by the function f or g, where f; g 2 C. The claim is that no algorithm has a distinguishing probability greater than m. Formally, let F and G be distributions over f0; 1g n f?1; +1g that produce hx; f (x a)i and hx; g(x a)i, respectively, where x is drawn according to the uniform distribution and a is drawn according to the noise distribution D. Also let F m and G m be induced distributions on m independent samples drawn according to F and G, respectively. We must show that there exists no prediction algorithm A (that outputs f0; 1g) with the property that for all f; g 2 C, Pr SF m[a(s) 1]? Pr SG m[a(s) 1] > m: Denote the above absolute difference of probabilities by A (f; g). Assume on the contrary that there exists an algorithm A such that A (f; g) > m. We will use the hybrid method or probability walk argument [4] to show that there is an algorithm B that can predict whether a labeled example hx; bi is drawn from F or from G with an advantage of 2 over random guessing, i.e., the algorithm B satisfies Pr[B(x; b) predicts correctly] > 1 2 + 2 : This is a contradiction by observing that the optimal Bayes predictor has a 2 advantage (see the appendix for details or [3]). Hence any algorithm needs m ((1?)) samples to distinguish any pair of -apart functions with probability at least 1?. We now elaborate on the hybrid method. Define a sequence of distributions H 0 ; H 1 ; : : : ; H n?1 over (f0; 1g n f?1; +1g) m where H i (hx 1 ; f (x 1 a)i; : : : ; hx i ; f (x i a)i; hx i+1 ; g(x i+1 a)i; : : : ; hx m ; g(x m a)i): Note that H 0 G m and H m F m. Now let p(h i ) Pr SHi [A(S) 1]. Then A (f; g) jp(h 0 )? p(h m )j m?1 i0 m?1 i0 (p(h i )? p(h i+1 )) jp(h i )? p(h i+1 )j So if A (f; g) > m then there exists i 0 2 f0; 1; : : : ; m?1g such that jp(h i0 )? p(h i0+1 )j > : Consider the following algorithm B: on input hx; bi, run algorithm A on the input Z (hy 1 ; f (y 1 a)i; : : : ; hy i0 ; f (y i0 a)i; hx; bi; hy i0+2 ; g(y i0+2 a)i; : : : ; hy m ; g(y m a)i) where y 1 ; : : : ; y i0 ; y i0+1 ; : : : ; y m are randomly drawn according to the uniform distribution. Assume without loss of generality that p(h i0 ) < p(h i0+1 ). So A is more likely to output 1 when the component (i 0 + 1) of its input is labeled by f than when it is labeled by g. Then B s prediction is given by f if A(Z) 1 B(hx; bi) g otherwise Let E B be the event that B makes a wrong prediction. The probability of event E B is given by (assuming a uniform prior on f and g, i.e., chosen uniformly) Pr[E B ] (Pr[E B j hx; bi F ] + Pr[E B j hx; bi G]) 1 2 1 2 [(1? p(h i0+1)) + p(h i0 )] 1 2? 1 2 [p(h i0+1)? p(h i0 )]: So algorithm B has an advantage strictly greater than 2 in prediction. 4.1 NEAR TIGHT CHARACTERIZATION In the following we will use Fourier analysis to obtain a nearly tight characterization of the noisy distance quantity D (f; g). Definition 3 Let f : f0; 1g n! f?1; +1g be a Boolean f0; 1gn function. Define the transformation T P, 2 [0; 1] on Boolean functions so that T (f ) c2f0;1g n ^f c (c) c. Then the -attenuated power spectrum of f is s (f ) jjt (f )jj 2 2 2 c ^f (c) 2 : As it turns out, D (f; g) is characterized by the -attenuated power spectrum of f? g with c E ad [ c (a)]. In particular, define s D (f ) to be s (f ) with c defined in this way. Then we have: Theorem 3 Let f; g : f0; 1g n! f?1; +1g be Boolean functions and D any probability distribution over f0; 1g n. Then s D (f? g) D (f; g) p s D (f? g): (1) Proof Using the fact that E[jj] p E[ 2 ], we get D (f; g) 1 2p ExUn [(E ad [f (x a)? g(x a)]) 2 ]: Let h(x) (f (x)? g(x))2. Then right hand side of the previous expression becomes q E x [E 2 a [h(x a)]]: c

We now work with the inner expression E x [E 2 a [h(x a)]]. E x [E 2 a [h(x a)] E x [E a [h(x a)]e b [h(x b)] E a;b [E x [ E a;b [ s s;t s D (h): Hence we get s;t ^h(s)^h(t) s (x a) t (x b)]] ^h(s)^h(t) s (a) t (b)e x [ s (x) t (x)] ^h(s) 2 E 2 a [ s(a)] D (f; g) p s D (f? g): Next, we show a lower bound on D (f; g). We note that 0 je a [h(x a)]j 1, since h 2 f?1; 0; +1g. Thus E x [je a [h(x a)]j] E x [E 2 a [h(x a)]] s D(h) using the same analysis as in the upper bound. This completes the theorem. Define SD (C) minfs D(f? g) j f; g 2 C with Pr[f 6 g] > g: U Using this definition with Theorem 3 we have the following inequalities. Theorem 4 For any class C and any we have S D (C) D (C) qs D (C): Then by Theorem 3 we have the following lower bound. Theorem 5 Let C be a concept class with SD (C) S. Then any PAC learning algorithm for C under D-distribution attribute noise that outputs an -good hypothesis with probability at least 1? requires a sample complexity of 1? p S. We now show that the class of parity functions is not PAC learnable under the uniform distribution with D-noise for almost every noise distribution D. Theorem 6 Let D be a distribution such that max x D(x) is superpolynomially small (or 1!(poly(n))). Then the set of parity functions is not PAC-learnable under D-distribution noise. Proof Notice that for any two distinct parity functions f and g we have Pr[f 6 g] 12. Since f and g are parity functions, s D (f? g) s D (f ) + s D (g), and it is enough to find two distinct parity functions f and g with superpolynomially small s D (f ) and s D (g). Consider E c [s D ( c )] where c is over the uniform distribution. We have E c [s D ( c )] E c [ 2 c ] E c [E ad [ c (a)]e bd [ c (b)]] E c E a;b [ c (a b)] E a;b E c [ ab (c)] E a;b I[a b] max D(x): x where I[a b] 1 if a b and 0 otherwise. Therefore, because s D (f ) is nonnegative for all D and Boolean f, only a superpolynomially small fraction of parity functions c can be inverse polynomially large if D(x) is superpolynomially small for all x. So there are at least two parity functions f and g for which both s D (f ) and s D (g) are superpolynomially small. Finally, it should be noted that Theorem 5 is only a hardness result for strong PAC learnability. As an example of a class that can be weakly learned in spite of arbitrary and unknown random attribute noise, consider monotone Boolean functions. Blum, Burch, and Langford [1] have shown that every monotone Boolean function is weakly approximated with respect to the uniform distribution by either one of the two constant functions or by the majority function. Since random attribute noise alone does not change the label of a function, it is easy to test noisy examples to see if one of the constants functions is a weak approximator to a montone function f. If not, majority is. 5 UPPER BOUNDS In this section we consider a certain type of Fourier-based learning algorithm which we will call LMN-style. The LMNstyle algorithm was introduced by Linial, Mansour, and Nisan [7], who showed that the class AC 0 of polynomial-size, constant depth circuits is PAC learnable with respect to the uniform distribution in quasipolynomial (roughly n polylog n ) time. The key to their result was analyzing the Fourier properties of AC 0 to show that for every AC 0 function f, the sum of the squares of the Fourier coefficients of degree polylog n or less is nearly 1. They then showed that this implied that the function h(x) sign 0 @ jajpolylog n ^f (a) a (x) 1 A is a good approximator to the target function f. Finally, it follows from standard Chernoff bounds that all of these Fourier coefficients can be closely approximated by sampling from a uniform-distribution example oracle, with sample size and running time dominated by n polylog n. An LMN-style algorithm, then, consists of estimating for every n-bit index in a set T Fourier coefficients, with the guarantee that the sum of the squares of these coefficients is nearly 1. How near 1 this sum must be, and therefore how large the set T must be, depends on the value of the PAC accuracy parameter, which is why subscripts T. For example, in the case of Linial et al. s algorithm for AC 0, the Hamming weight of the Fourier indices grows as approaches 0. The hypothesis resulting from an LMN-style algorithm will be of the form h(x) sign a2t ~ f (a) a (x) where ~ f (a) represents an estimate of the Fourier coefficient ^f (a). In this section we show that if there is an LMN-style algorithm for learning a class of functions C, then C is PAClearnable under any (D; R)-noise in time polynomial in jt j,! ;

1(1? 2), and 1, where is the expectation of the noise rate in the label. Since 1 is a lower bound for PAClearning with D-distribution noise and 1(1? 2) is a lower bound for learning with label noise [9], our result is tight (up to polynomial factors). We now formally state the result: Theorem 7 Let C be a class of Boolean functions and suppose that C is learnable with respect to the uniform distribution by an LMN-style algorithm using index set T. Then for every such that the set of parity functions indexed by T are a subset of C, C is learnable with respect to the uniform distribution and with any known (D; R)-noise in time polynomial in 1; 1; 1(1? 2); jt O() j, and 1 D. For the LMN-style algorithm for AC 0, as long as 1 O(n polylog n ), the parity functions indexed by T are in AC 0 (by results in Linial et al. [7] and the fact that parity on polylogarithmic bits can be computed in AC 0 ). This gives us immediately: Theorem 8 For 1 O(n polylog n ), the class AC 0 of constant depth, polynomial size circuits is learnable under the uniform distribution with any known (D; R)-noise in time dominated by n poly(log n) poly(1 D ): As a specific example of the application of this theorem, we claim that if the attribute noise rate on all attributes is independent with rate O(1polylog n) for each attribute (but possibly different values for each) then there is a learning algorithm for AC 0 with time dominated by n polylog n. To see this, recall that the hypothesis in an LMN-style algorithm is formed using only (estimates of) coefficients indexed by T, and that for AC 0 all of these indices have polylogarithmic Hamming weight. Furthermore, based on results of Linial et al. [7], if f and g are Boolean functions such that Pr[f 6 g] > (1n polylog n ) then the difference ^f(c)? ^g(c) must be at least 1n polylog n large for at least one of the coefficients indexed by T. But then s D (f? g) P c 2 c d f? g 2 (c) P c 2 c ( ^f (c)? ^g(c)) 2 (the final equality follows by linearity of the Fourier transform) will be inverse quasipolynomially large as along as c E ad [ c (a)] is inverse quasipolynomial for all c in T. But then a simple probabilistic analysis shows that in fact all of these c will be sufficiently large as long as jcj is polylogarithmic in n. In particular, c E ad [(?1)P n i1 aici ] ny i1 Y i2c E aidi [(?1) aici ]; D is product (1? 2p i ) > (1? 1poly(log n)) jcj ; 8i,p i < 1poly(log n) > 1n poly(log n) ; since jcj poly(log n) Therefore, D is inverse quasipolynomially large, and our claim follows by the theorem. Sketch of the Proof of Theorem 7: Let D (C) and T T. First we show that there is at most one c 2 T that is less than 2. Suppose for the contrary there are two c1 and c2 such that Then c1 < 2; c2 < 2: 2 s D ( c1? c2 ) s D ( c1 ) + s D ( c2 ) 2 c1 + 2 c2 2 2 and we have a contradiction. Now to find the coefficient of c 2 T we take a sample S f(x i a i ; f (x i )b i )g (recall that since the function f is f+1;?1g-valued we choose b i 2 f?1; +1g, so OR becomes multiplication) and estimate the expectation E x;a;b [f (x)b c (x a)]. Now E x;a;b [f (x)b c (x a)] E b [b]e x [f (x) c (x)]e a [ c (a)] (1? 2) ^f (c) c : Because we are assuming that D and R are known, the factors of (1? 2) and c are easily eliminated for those c s such that c 2. Thus, for such c s, a good estimate of the above expectation gives a good estimate of the Fourier coefficient ^f(c). And by Chernoff, to estimate this expectation we only need sample size and time poly 1 ^f(c) c (1? 2) Finally, we can ignore c s such that ^f (c) < p 2jT j, since this will still give us a sum of squares of coefficients very near 1. This shows that all of the relevant coefficients indexed by T, except maybe one (call it c 0 ), can be estimated in time polynomial in 1poly(jT j) and 1. To estimate the coefficient of ^f (c 0 ) (only an approximation of this coefficient is required) we can use the fact that the sum of all of the relevant coefficients indexed by T should be nearly 1. If after we have found estimates for all the coefficients for indices other than c 0 the resulting sum of squares is noticeably less than 1, then we know that ^f 2 (c 0 ) must be approximately the difference. Thus we can estimate ^f(c 0 ) as well, although perhaps not quite as accurately as the other coefficients. However, given the form of the LMN-style hypothesis h, it can be argued that this estimate can be made sufficiently close to ensure that h is an -approximator. Full details will be included in the final paper. However, we note that learning even very simple classes of functions can be hard if the attribute noise D is unknown. Claim 9 There exists no uniform-distribution learning algorithm for the class f e1 ; e2 g under an unknown attribute noise D, i.e., (D; 0)-noise. Proof Let D 1 ; D 2 be distributions over f0; 1g 2 defined as follows: D 1 is the product noise distribution (12; 14), i.e.,! :

attribute x 1 is simply uniform random noise, attribute x 2 has noise rate 14, and no other attributes have noise applied. Similarly, let D 2 be the product noise distribution (14; 12). Clearly, no algorithm can distinguish between learning e1 under D 1 -noise and learning e2 under D 2 - noise, for if a D 1 and b D 2 then the random variables (x; e2 (x a)); are identically distributed. (x; e1 (x b)) f and g. Note that jp b f (x)?pb g (x)j je a[(f?g)(xa)]j2. Now allowing x to vary according to some distribution D, the expected loss of the Bayes predictor is L 1 2? 1 2 E x[jp b f (x)? pb g (x)j] 1 2? 1 2 D(f; g): By the same token, there is no predictor that has smaller expected loss than L. References [1] Avrim Blum, Carl Burch, and John Langford. On Learning Monotone Boolean Functions. In Proceedings of the 39th Annual IEEE Symposium on Foundations of Computer Science, 1998. [2] Scott Decatur and Rossario Gennaro. On Learning from Noisy and Incomplete Examples. In Proceedings of the 8th Annual ACM Conference on Computational Learning Theory, 353 360, 1995. [3] Luc Devroye, László Györfi, and Gábor Lugosi. A Probabilistic Theory of Pattern Recognition. Springer- Verlag, 1996. [4] Shafi Goldwasser and Silvio Micali. Probabilistic Encryption. In Journal of Computer and System Sciences, 28(2):270-299, 1984. [5] Sally Goldman and Robert Sloan. Can PAC Learning Algorithms Tolerate Random Attribute Noise? In Algorithmica, 14(1):70-84, 1995. [6] M. Kearns and M. Li. Learning in the Presence of Malicious Errors. In SIAM Journal on Computing, 22(4):807 837, 1993. [7] Nathan Linial, Yishay Mansour, and Noam Nisan. Constant Depth Circuits, AC 0 Circuits, and Learnability. In Journal of the ACM, 40(3):607-620, 1993. [8] George Shackelford and Dennis Volper. Learning k- DNF with Noise in the Attributes. In Proceedings of the 1988 Workshop on Computational Learning Theory, 97-103, 1988. [9] Hans Ulrich Simon. General Bounds on the Number of Examples Needed for Learning Probabilistic Concepts. In Proceedings of the 6th Annual ACM Workshop on Computational Learning Theory, pages 402-411, 1993. APPENDI A We prove the connection between D (f; g) and the optimal Bayes loss. Consider a fixed x and a predictor that is given a bit b that is either f (x a) or g(x a), for a randomly chosen a from a simple product distribution with rate p, and must predict whether b is labeled according to f or g. For b 2 f?1; +1g, let p b f (x) Pr[f (x a) b] and pb g (x) Pr[g(xa) b]. Then, by standard Bayes theory, the Bayes predictor B, whose decision is given by f B(hx; bi) g if p b f (x) > p b g (x) otherwise achieves a discrete loss of 12? jp b f (x)? pb g (x)j2, for the given fixed x, and this quantity is the smallest achievable loss over all predictors. Here we are assuming uniform priors on