27
How Robust are Linear Sketches to Adaptive Inputs? Moritz Hardt, David P. Woodruff IBM Research Almaden

How Robust are Linear Sketches to Adaptive Inputs?

  • Upload
    lysa

  • View
    38

  • Download
    0

Embed Size (px)

DESCRIPTION

How Robust are Linear Sketches to Adaptive Inputs?. Moritz Hardt , David P. Woodruff IBM Research Almaden. Two Aspects of Coping with Big Data. Efficiency Handle enormous inputs. Robustness Handle adverse conditions. Big Question: Can we have both?. Sketch y = Ax in R r. - PowerPoint PPT Presentation

Citation preview

Page 1: How Robust are Linear Sketches to Adaptive Inputs?

How Robust are Linear Sketches to Adaptive Inputs?

Moritz Hardt, David P. WoodruffIBM Research Almaden

Page 2: How Robust are Linear Sketches to Adaptive Inputs?

Two Aspects of Coping with Big Data

EfficiencyHandle

enormous inputs

RobustnessHandle

adverse conditions

Big Question: Can we have both?

Page 3: How Robust are Linear Sketches to Adaptive Inputs?

Algorithmic paradigm: Linear Sketches

Linear map A

Applications: Compressed sensing, data streams, distributed computation, ...

Unifying idea: Small number of linear measurements applied to data

Data vector x in Rn

Output Sketch y = Ax in Rr

r << n For this talk: output can be any (not necessarily efficient) function of y

Page 4: How Robust are Linear Sketches to Adaptive Inputs?

“For each” correctnessFor each x: Pr { Alg(x) correct } > 1 – 1/poly(n)

Pr over randomly chosen matrix A

Does this imply correctness on many inputs?

No guarantee if input x2 depends on Alg(x1) for earlier input x1

Only under modeling assumption:Inputs are non-adaptively chosen

Why not?

Page 5: How Robust are Linear Sketches to Adaptive Inputs?

Example: Johnson-Lindenstrauss Sketch• Goal: estimate |x|2 from |Ax|2

• JL Sketch: if A is a k x n matrix of i.i.d. N(0, 1/k) random variable with k > log n, then Pr[|Ax|2 = (1±1/2)|x|2] > 1-1/poly(n)

• Attack: 1. Query x = ei and x = ei + ej for all standard unit vectors ei and ej

• Learn |Ai|2, |Aj|2, |Ai + Aj|2, so learn <Ai, Aj>

2. Hence, learn AT A, and learn kernel of A

3. Query a vector x 2 kernel(A)

Page 6: How Robust are Linear Sketches to Adaptive Inputs?

Benign/NaturalMonitor traffic using sketch, re-route traffic based on output, affects future inputs.

Adversarial DoS attack on network monitoring unit

Correlations arise in nearly any realistic setting

In this work: Broad impossibility results

Can we thwart the attack?

Can we prove correctness?

Page 7: How Robust are Linear Sketches to Adaptive Inputs?

Benchmark Problem

GapNorm(B): Given decide if(YES)(NO)

Easily solvable for B = 1+ε using “for each” guarantee by sketch with O(log n/ε2) rows using JL.

Goal: Show impossibility for very basic problem.

Page 8: How Robust are Linear Sketches to Adaptive Inputs?

Main ResultTheorem. For every B, given oracle access to a linear sketch using dimension r · n – log(Bn), we can find in time poly(r,B) a distribution over inputs on which sketch fails to solve GapNorm(B)

Corollary. Same result for any lp-norm.Corollary. Same result even if algorithm uses internal randomness on each query.

Efficient attack (rules out crypto), even slightlynon-trivial sketching dimension impossible

Page 9: How Robust are Linear Sketches to Adaptive Inputs?

Application to Compressed Sensing

Theorem. No linear sketch with o(n/C2) rows gurantees l2/l2 sparse recovery with approximation factor C on a polynomial number of adaptively chosen inputs.

l2/l2 recovery: on input x, output x’ for which:

Note: Impossible to achieve with deterministic matrix A, but possible with “for each” guarantee with r = k log(n/k).

[Gilbert-Hemenway-Strauss-W-Wootters12] has some positive results

Page 10: How Robust are Linear Sketches to Adaptive Inputs?

Outline

• Proof of Main Theorem for GapNorm– Proved using “Reconstruction Attack”

• Sparse Recovery Result– By Reduction from GapNorm– Not in this talk

Page 11: How Robust are Linear Sketches to Adaptive Inputs?

Computational Model

Definition (Sketch)An r-dimensional sketch is anyfunction satisfying

for some subspace

Sketch has unbounded computational poweron top of PUx

1. Sketches Ax and UTx are equivalent, where UT has orthonormal rows and row-span(UT) = row-span(A)

2. Sketch UTx equivalent to PU x = UUT x

Why?

Page 12: How Robust are Linear Sketches to Adaptive Inputs?

Algorithm (Reconstruction Attack)Input: Oracle access to sketch f using unknown subspace U of dimension rPut V0 = {0}, subspace of 0 dimensionFor t = 1 to t = r:

(Correlation Finding) Find vectors x1,...,xm weakly correlated with unknown subspace U, orthogonal to Vt-1(Boosting) Find single vector x strongly

correlated with U, orthogonal to Vt-1(Progress) Put Vt = span{Vt-1, x}

Output: Subspace Vr

Page 13: How Robust are Linear Sketches to Adaptive Inputs?

Algorithm (Reconstruction Attack)Input: Oracle access to sketch f using unknown subspace U of dimension rPut V0 = {0}, empty subspaceFor t = 1 to t = r:

(Correlation Finding) Find vectors x1,...,xm weakly correlated with unknown subspace U, orthogonal to Vt-1

(Boosting) Find single vector x strongly correlated with U, orthogonal to Vt-1

(Progress) Put Vt = Vt-1 + span{x}Output: Subspace Vr

Page 14: How Robust are Linear Sketches to Adaptive Inputs?

Conditional Expectation Lemma

Moreover, 1. Δ ≥ poly(1/d)2. g = N(0,σ)n for a carefully chosen σ unknown to sketching algorithm

Lemma. Given d-dimensional sketch f, we can find using poly(d) queries adistribution g such that:

“Advantage over random”

Page 15: How Robust are Linear Sketches to Adaptive Inputs?

SimplificationFact: If g is Gaussian, then PUg = UUTg is Gaussian as well

Hence, can think of query distribution as choosing random Gaussian g to be inside subspace U.

We drop the PU projection operator for notational simplicity.

Page 16: How Robust are Linear Sketches to Adaptive Inputs?

The three step intuition

(Symmetry) Since the queries are random Gaussian inputs g with an unknown variance, by spherical symmetry, sketch f learns nothing more about query distribution than norm |g|(Averaging) If |g| is larger than expected, the sketch is “more likely” to output 1(Bayes) Hence, by sort-of-Bayes-Rule, conditioned on f(g)=1, expectation of |g| is likely to be larger

Page 17: How Robust are Linear Sketches to Adaptive Inputs?

Def. Let p(y) = Pr{ f(y) = 1 }y in U uniformly random with |y|2 = s

Fact. If g is Gaussian with E|g|2 = t, then,

density of χ2-distribution with expectation t

and d degrees of freedom

Page 18: How Robust are Linear Sketches to Adaptive Inputs?

p(s) = Pr(f(y) = 1) y in U unif. random with |y|2 = s

?

Norm s

1

0d/n Bd/n

By correctness of sketch

l r

Page 19: How Robust are Linear Sketches to Adaptive Inputs?

Sliding χ2-distributions

• ¢(s) = sl r (s-t) vt(s) dt

• ¢(s) < 0 unless s > r – O(r1/2 log r)• s0

1 ¢(s) ds =s01 sl

r (s-t) vt(s) dt ds = 0

Page 20: How Robust are Linear Sketches to Adaptive Inputs?

Averaging Argument

• h(s) = E[f(gt) | |g|2 = s]

• Correctness:– For small s, h(s) ¼ 0, while for large s, h(s) ¼ 1

• s01 h(s) ¢(s) ds =s0

1 sl r h(s) (s-t) vt(s) dt ds ¸ d

• 9 t so that s01 h(s) s vt(s) ds ¸ t s0

1 vt(s) ds + d/(l-r)

• For this t, E[|gt|2 | f(gt) = 1] ¸ t + ¢

Page 21: How Robust are Linear Sketches to Adaptive Inputs?

Algorithm (Reconstruction Attack)Input: Oracle access to sketch f using unknown subspace U of dimension rPut V0 = {0}, empty subspaceFor t = 1 to t = r:

(Correlation Finding) Find vectors x1,...,xm weakly correlated with unknown subspace U, orthogonal to Vt-1

(Boosting) Find single vector x strongly correlated with U, orthogonal to Vt-1

(Progress) Put Vt = Vt-1 + span{x}Output: Subspace Vr

Page 22: How Robust are Linear Sketches to Adaptive Inputs?

Boosting small correlations

x1

x2

x3

.

.

.

xm

n

poly(r)

M =

1. Sample poly(r) vectors using CoEx Lemma

2. Compute top singular vector x of M

Lemma: |PUx| > 1-poly(1/r)

Proof: Discretization +Concentration

Page 23: How Robust are Linear Sketches to Adaptive Inputs?

Implementation in poly(r) time

• W.l.og. can assume n = r + O(log nB)

– Restrict host space to first r + O(log nB)

coordinates

• Matrix M is now O(r) x poly(r)

• Singular vector computation poly(r) time

Page 24: How Robust are Linear Sketches to Adaptive Inputs?

Iterating previous steps

Generalize Gaussian to “subspace Gaussian” = Gaussian vanishing on maintained subspace Vt

Intuition: Each step reduces sketch dimension by one.

After r steps:1. Sketch has no dimensions left!2. Host space still has n – r > O(log nB) dimensions

Page 25: How Robust are Linear Sketches to Adaptive Inputs?

Problem

Top singular vector not exactly contained in UFormally, sketch still has dimension r

Can fix this by adding small amount of Gaussiannoise to all coordinates

Page 26: How Robust are Linear Sketches to Adaptive Inputs?

Algorithm (Reconstruction Attack)Input: Oracle access to sketch f using unknown subspace U of dimension rPut V0 = {0}, empty subspaceFor t = 1 to t = r:

(Correlation Finding) Find vectors x1,...,xm weakly correlated with unknown subspace U, orthogonal to Vt-1

(Boosting) Find single vector x strongly correlated with U, orthogonal to Vt-1

(Progress) Put Vt = Vt-1 + span{x}Output: Subspace Vr

Page 27: How Robust are Linear Sketches to Adaptive Inputs?

Open Problems

• Achievable polynomial dependence still open

• Optimizing efficiency alone may lead to non-robust algorithms - What is the trade-off between robustness

and efficiency in various settings of data analysis?