1 Formulating The Low Degree Testing Problem



Similar documents
1 The Line vs Point Test

Linearity and Group Homomorphism Testing / Testing Hadamard Codes

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

minimal polyonomial Example

2.1 Complexity Classes

(67902) Topics in Theory and Complexity Nov 2, Lecture 7

The Mean Value Theorem

Lecture 4: AC 0 lower bounds and pseudorandomness

Factoring & Primality

FIRST YEAR CALCULUS. Chapter 7 CONTINUITY. It is a parabola, and we can draw this parabola without lifting our pencil from the paper.

FACTORING POLYNOMIALS IN THE RING OF FORMAL POWER SERIES OVER Z

Notes 11: List Decoding Folded Reed-Solomon Codes

A Brief Introduction to Property Testing

The Epsilon-Delta Limit Definition:

Modern Algebra Lecture Notes: Rings and fields set 4 (Revision 2)

Applied Algorithm Design Lecture 5

Prime Numbers and Irreducible Polynomials

Homework # 3 Solutions

Influences in low-degree polynomials

The finite field with 2 elements The simplest finite field is

Lecture 3: Finding integer solutions to systems of linear equations

H/wk 13, Solutions to selected problems

15. Symmetric polynomials

Chapter 4, Arithmetic in F [x] Polynomial arithmetic and the division algorithm.

Zeros of Polynomial Functions

Practice with Proofs

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs

Linear Codes. Chapter Basics

Continued Fractions and the Euclidean Algorithm

1 Lecture: Integration of rational functions by decomposition

3.2 The Factor Theorem and The Remainder Theorem

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

Revised Version of Chapter 23. We learned long ago how to solve linear congruences. ax c (mod m)

MOP 2007 Black Group Integer Polynomials Yufei Zhao. Integer Polynomials. June 29, 2007 Yufei Zhao

Lecture 1: Course overview, circuits, and formulas

On the Unique Games Conjecture

Factoring Cubic Polynomials

1 if 1 x 0 1 if 0 x 1

Algorithmic and Analysis Techniques in Property Testing

MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform

it is easy to see that α = a

The Division Algorithm for Polynomials Handout Monday March 5, 2012

Lecture 10: Distinct Degree Factoring

7. Some irreducible polynomials

Factorization Algorithms for Polynomials over Finite Fields

Lectures 5-6: Taylor Series

Math 4310 Handout - Quotient Vector Spaces

(a) Write each of p and q as a polynomial in x with coefficients in Z[y, z]. deg(p) = 7 deg(q) = 9

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include

Inner Product Spaces

6 EXTENDING ALGEBRA. 6.0 Introduction. 6.1 The cubic equation. Objectives

NP-Completeness I. Lecture Overview Introduction: Reduction and Expressiveness

2.3 Convex Constrained Optimization Problems

Quotient Rings and Field Extensions

Tim Kerins. Leaving Certificate Honours Maths - Algebra. Tim Kerins. the date

1 Message Authentication

5. Factoring by the QF method

Adaptive Online Gradient Descent

Universal Hash Proofs and a Paradigm for Adaptive Chosen Ciphertext Secure Public-Key Encryption

U.C. Berkeley CS276: Cryptography Handout 0.1 Luca Trevisan January, Notes on Algebra

Some facts about polynomials modulo m (Full proof of the Fingerprinting Theorem)

Math 231b Lecture 35. G. Quick

ON GALOIS REALIZATIONS OF THE 2-COVERABLE SYMMETRIC AND ALTERNATING GROUPS

Approximation Algorithms

Lecture 11: The Goldreich-Levin Theorem

Notes from Week 1: Algorithms for sequential prediction

Basics of Polynomial Theory

How To Prove The Dirichlet Unit Theorem

FOUNDATIONS OF ALGEBRAIC GEOMETRY CLASS 22

FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES

CHAPTER SIX IRREDUCIBILITY AND FACTORIZATION 1. BASIC DIVISIBILITY THEORY

Private Approximation of Clustering and Vertex Cover

PROBLEM SET 6: POLYNOMIALS

Breaking Generalized Diffie-Hellman Modulo a Composite is no Easier than Factoring

Short Programs for functions on Curves

Integer Factorization using the Quadratic Sieve

Section 3.7. Rolle s Theorem and the Mean Value Theorem. Difference Equations to Differential Equations

1 Homework 1. [p 0 q i+j p i 1 q j+1 ] + [p i q j ] + [p i+1 q j p i+j q 0 ]

3 1. Note that all cubes solve it; therefore, there are no more

LINEAR ALGEBRA W W L CHEN

Integer roots of quadratic and cubic polynomials with integer coefficients

Chapter 20. Vector Spaces and Bases

Lecture 13 - Basic Number Theory.

Math 55: Discrete Mathematics

OHJ-2306 Introduction to Theoretical Computer Science, Fall

Homework until Test #2

0.1 Phase Estimation Technique

Zeros of a Polynomial Function

Offline sorting buffers on Line

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

Recursive Estimation

A Sublinear Bipartiteness Tester for Bounded Degree Graphs

Lecture 14: Section 3.3

Polynomial Degree and Lower Bounds in Quantum Complexity: Collision and Element Distinctness with Small Range

CSC2420 Fall 2012: Algorithm Design, Analysis and Theory

Introduction to Algebraic Geometry. Bézout s Theorem and Inflection Points

MATH 425, PRACTICE FINAL EXAM SOLUTIONS.

Vieta s Formulas and the Identity Theorem

Lecture 2: Universality

Rolle s Theorem. q( x) = 1

Transcription:

6.895 PCP and Hardness of Approximation MIT, Fall 2010 Lecture 5: Linearity Testing Lecturer: Dana Moshkovitz Scribe: Gregory Minton and Dana Moshkovitz In the last lecture, we proved a weak PCP Theorem, namely, NP P CP 1,0.99 [O(log n), polylog(n)], under the low degree testing assumption, i.e., the assumption that given the table of a function f : F m F, a degree d and access to an auxiliary proof, a probabilistic verifier can test whether f corresponds to a polynomial of degree at most d using a poly-logarithmic number of queries. In this lecture we discuss what should be the exact formulation of the low degree testing assumption. Then we prove this formulation for the special case of d = 1 and F = GF (2). This is the Blum-Luby-Rubinfeld linearity test. In the next lecture we deal with the harder case of general degree d and field F. 1 Formulating The Low Degree Testing Problem Can a local algorithm distinguish polynomials of degree at most d from functions that are not polynomials of degree at most d? The answer to this question is no. For let p be a polynomial of degree at most d, and let p be the function { p p(x) if x x0 (x) = p(x 0 ) + 1 if x = x 0. The function p p is not the zero function, but it has a root everywhere except one point. Because a nonzero polynomials of degree at most d have roots in at most d/ F fraction of the space, we deduce that p p, and so p, cannot be polynomials of degree at most d. Thus a verifier which purports to test for low degree polynomials should pass p but fail p. But of course the probability that any verifier can distinguish p and p is bounded above by the probability that it queries location x 0! If the verifier makes few queries, and we choose x 0 appropriately, then this probability can be made quite small. (To be more precise, if the proof has length l and the verifier makes q queries, then there exists some x 0 for which the probability that the verifier queries x 0 is q/l. We are interested in verifiers in which the proof length grows faster than the number of queries, so this is o(1).) With this in mind, we cannot hope for an exact test of low degree polynomials. What we ask for instead is testing whether the function is close to a low degree polynomial in Hamming distance. Definition 1 (Far/close to low degree). We say that a function f : F m F is δ-far from degree d if for any m-variate polynomial p of degree at most d over F, we have (f, p) δ where we view f and p as vectors (so (f, p) = Pr x F m [f(x) p(x)]). We say that f is δ-close to degree d, if there exists an m-variate polynomial p of degree at most d over F, such that (f, p) δ. 1

2 Note that, by definition, if we query a function that is (1 γ)-close to a low degree polynomial on a uniformly random point, then with probability at least (1 γ) we get the same answer as if we queried the low degree polynomial. We will use this observation (as well as additional ideas) to argue that the low degree testing assumption we formulate suffices to prove the weak PCP Theorem. Assumption 1.1 (Low Degree Testing). There are constants δ, γ, γ > 0, such that given the table of a function f : F m F and a degree d δ F, there is a probabilistic verifier for the statement deg f d. The verifier is given access to f and to an auxiliary proof and satisfies the following: Completeness: If deg f d, then there is a proof that the verifier always accepts. Soundness: If f is γ-far from degree d, then for any proof, the verifier rejects with probability at least γ. The verifier uses O(log( F m )) random bits. It makes only F O(1) queries to f. Put differently, the soundness condition says that if, for some auxiliary proof, the verifier accepts with probability more than 1 γ, then there is a polynomial p of degree at most d that agrees with f on 1 γ fraction of the points in F m. 2 Linear Functions In this lecture we consider the case of d = 1, F = GF (2). That is, given a function f : {0, 1} n {0, 1}, we want to test (with high probability) whether f is (close to) a linear function. This special case is not sufficient for proving the PCP Theorem, but it is simpler and its ideas important. We first define what we mean by a linear function. Two different definitions come to mind, so let us show that they are equivalent. Definition. A function f : {0, 1} n {0, 1} is linear if for all x, y {0, 1} n, f(x) + f(y) = f(x + y). Claim. A function f : {0, 1} n {0, 1} is linear iff there exists a vector a = (a 1,..., a n ) {0, 1} n such that f(x) = n i=1 a ix i = a, x. (As we are working in GF (2), addition is of course taken modulo 2.) Proof. The ( ) direction is clear. For the ( ) direction, let {e 1,..., e n } be the standard basis, i.e. e i = (0,..., 0, 1, 0,..., 0), } {{ } 1 in position i and define a i = f(e i ). Then the stated formula follows by linearity. Note that the functions we are calling linear are not affine functions y = ax + b; we require that the constant term be zero.

3 The Blum-Luby-Rubinfeld Linearity Test 3 The test we show does not use an auxiliary proof; it only makes queries to f. In this case, the tester has to make at least three queries, as there exist functions f which are far from linear but which have the property that for any two points x, y, there exists a linear function g such that f(x) = g(x) and f(y) = g(y). We will show that three queries suffices, by studying the properties of the following simple test. BLR Test (Blum, Luby, Rubinfeld). Choose uniformly random points x, y {0, 1} n. Test if f(x) + f(y) = f(x + y). This algorithm uses 2n random bits. It makes q = 3 queries. The completeness of this test is 1, because obviously a linear function passes with probability 1. Analyzing the soundness is the interesting part; the answer is given by the following theorem. Theorem (Soundness of BLR). If f is δ-far from linear, then ( 2 Pr[BLR test rejects f] min 9, δ ) 2 Before proving this theorem, we make some remarks. 2δ 9. This result is not tight, as we can prove via Fourier analysis that Pr[rejection] δ. (We will return to this later in the course.) And even that result is not tight in the low order terms! The definition of linearity generalizes to any group G; in the setting of group theory such a map is known as a homomorphism. In fact, the BLR test generalizes to testing homomorphisms on groups. In this setting, the soundness theorem above is tight. For instance, define f : (Z/9) n Z/9 by f(u) = 3k if u 1 {3k 1, 3k, 3k + 1}. (That is, f rounds the first coordinate to the nearest multiple of 3.) This is not linear; one can check that f is (2/3)-far from linear. Hence the soundness theorem tells us that BLR should reject f with probability at least 2/9. In fact, that is exactly the rejection probability, because f(x) + f(y) f(x + y) x 1 y 1 1 mod 3 or x 1 y 1 1 mod 3. 4 Analysis of the Linearity Test The rest of this lecture is devoted to proving the soundness theorem. 4.1 Majority Correction The proof uses the useful idea of majority correction. Fix a function f : {0, 1} n {0, 1} and a point x {0, 1} n. If f is linear, then for any y {0, 1} n we have f(x) = f(y) + f(x y). Thus we may think of each of the 2 n values of y as offering the vote f(y) + f(x y) for f(x). As there are only two possible values for f(x), 0 and 1, one of them must get a majority of the votes. We define a function g by setting g(x) to be the value that receives the most votes.

More formally, g : {0, 1} n {0, 1} is defined by 4 { 1 if Pry [f(y) + f(x y) = 1] 1/2 g(x) = 0 otherwise. (We have chosen to always break a tie with the value 1; this was arbitrary, and it will turn out that the definition of g(x) in the case Pr y [f(y) + f(x y) = 1] = 1/2 is unimportant.) It will be useful later to define the probabilities By definition of g(x), P x 1/2 for all x. 4.2 Majority Correction Works P x = Pr y [g(x) = f(y) + f(x y)]. We will obtain the soundness theorem by proving three claims relating properties of the BLR test to the function g. Claim. Pr[BLR rejects f] 1 2 dist(g, f). Proof. Conditioning on whether g(x) = f(x) or not, we have Pr[rejection] = Pr[g(x) f(x)] Pr[rejection g(x) f(x)] + Pr[g(x) = f(x)] Pr[rejection g(x) = f(x)]. We get a lower bound on Pr[rejection] by ignoring the second term. In the first term, notice that Pr[g(x) f(x)] = dist(g, f) by definition of the distance. By definition of g, if g(x) f(x) then f(x) = f(y) + f(x y) for 1 P x 1/2 of the possible values of y. But because we are working over the binary field (so addition and subtraction are the same), the equation f(x) = f(y)+f(x y) is equivalent to the BLR test f(x+y) = f(x)+f(y). Hence, given g(x) f(x), the BLR test fails with probability at least 1/2. Putting this together, Pr[rejection] 1 2 dist(g, f), as desired. Claim. If Pr[BLR rejects f] < 2 9, then for all x we have P x > 2 3. Proof. Fix x. We compute A = Pr y,z [f(y) + f(x + y) = f(z) + f(x + z)] in two different ways. First, notice that f(y) + f(x + y) equals g(x) with probability P x, and the same is true of f(z) + f(x + z). Using independence of y and z, the probability that both expressions are equal to g(x) is P 2 x, and the probability that they are both equal to g(x) + 1 is (1 P x ) 2. Hence A = P 2 x + (1 P x ) 2. We can also bound A using the probability of BLR rejection. First, rewrite the condition f(y) + f(x + y) = f(z) + f(x + z) as f(y) + f(z) = f(x + y) + f(x + z). By definition of the BLR test, f(y) + f(z) equals f(y + z) with probability 1 Pr[BLR rejects f] > 7/9. As y and z are independent and uniformly sampled, the same is true of x + y and x + z, and so the same argument shows that f(x + y) + f(x + z) = f((x + y) + (x + z)) = f(y + z) with probability > 7/9. Thus f(y) + f(z) = f(y + z) = f(x + y) + f(x + z)

with probability > 5/9, so certainly A = Pr[f(y) + f(z) = f(x + y) + f(x + z)] > 5/9. Combining the results of the last two paragraphs, we deduce that 5 P 2 x + (1 P x ) 2 > 5/9. This implies either P x < 1/3 or P x > 2/3. Of course the first case is impossible because P x 1/2, so we must have P x > 2/3 as desired. Claim. If Pr[BLR rejects f] < 2 9, then g is linear. Proof. By the previous claim, we must have P x > 2/3 for all x. Now fix x, y and consider choosing z uniformly at random. Then g(x) equals f(z) + f(x + z) with probability larger than 2/3. Similarly, g(y) equals f(z)+f(y +z) with probability larger than 2/3. The same argument says that g(x + y) equals f(z) + f(z + x + y) with probability larger than 2/3; we can of course replace the uniformly sampled value z by z + x, finding that g(x + y) = f(z + x) + f(z + y) with the same probability (more than 2/3). As each of these three conditions holds with probability larger than 2/3, they hold simultaneously with positive probability. That is, there exists at least one z 0 such that all hold. But this shows that g(x) = f(z 0 ) + f(x + z 0 ), g(y) = f(z 0 ) + f(y + z 0 ), and g(x + y) = f(z 0 + x) + f(z 0 + y) g(x) + g(y) = f(z 0 ) + f(x + z 0 ) + f(z 0 ) + f(y + z 0 ) = f(x + z 0 ) + f(y + z 0 ) = g(x + y). This holds for all x, y, so g is linear, as desired. Putting the last few claims together, we immediately get the soundness theorem. Specifically, we find that either Pr[rejection] 2/9, or else g is linear and so Pr[rejection] 1 2 dist(g, f) 1 dist(f, linear). 2 That is exactly what the soundness theorem asserts, so we are done.