Deterministic Discrepancy Minimization

Post on 24-Feb-2016

49 views 0 download

description

Deterministic Discrepancy Minimization. Nikhil Bansal (TU Eindhoven) Joel Spencer (NYU). S 3. S 4. S 1. S 2. Combinatorial Discrepancy. Universe: U= [1,…,n] Subsets: S 1 ,S 2 ,…, S m Problem: Color elements red/blue so each subset is colored as evenly as possible. - PowerPoint PPT Presentation

transcript

1/17

Deterministic Discrepancy Minimization

Nikhil Bansal (TU Eindhoven)Joel Spencer (NYU)

2/17

Combinatorial DiscrepancyUniverse: U= [1,…,n] Subsets: S1,S2,…,Sm

Problem: Color elements red/blue so each subset is colored as evenly as possible.

CS: Computational Geometry, Comb. Optimization, Monte-Carlo simulation, Machine learning, Complexity, Pseudo-Randomness, …Math: Dynamical Systems, Combinatorics, Mathematical Finance, Number Theory, Ramsey Theory, Algebra, Measure Theory, …

S1

S2

S3

S4

3/17

General Set System

Universe: U= [1,…,n] Subsets: S1,S2,…,Sm

Find : [n] ! {-1,+1} toMinimize |(S)|1 = maxS | i 2 S (i) |

For simplicity consider m=n henceforth.

4/17

Simple AlgorithmRandom: Color each element i independently as x(i) = +1 or -1 with probability ½ each.

Thm: Discrepancy = O (n log n)1/2

Pf: For each set, expect O(n1/2) discrepancyStandard tail bounds: Pr[ | i 2 S x(i) | ¸ c n1/2 ] ¼ e-c2

Union bound + Choose c ¼ (log n)1/2

Analysis tight: Random actually incurs ((n log n)1/2).

5/17

Better Colorings Exist![Spencer 85]: (Six standard deviations suffice) Always exists coloring with discrepancy · 6n1/2

Tight: Cannot beat 0.5 n1/2 (Hadamard Matrix, “orthogonal” sets)

Inherently non-constructive proof (pigeonhole principle on exponentially large universe)

Challenge: Can we find it algorithmically ? (Certain algorithms do not work)

Conjecture [Alon-Spencer]: May not be possible.

6/17

Algorithmic Results

[Bansal 10]: Efficient (randomized) algorithm for Spencer’s result.

Technique: SDPs (new rounding idea) Use several SDPs over time (guided by the non-constructive method).

General: Geometric problems, Beck Fiala setting, k-permutation problem, pseudo-approximation of discrepancy, …

Thm: Deterministic Algorithm for Spencer’s (and other) results.

This Talk•  

A

𝑥1… 𝑥𝑛Goal: Round to -1 or 1 Minimize error for each row

Chernoff: Error = Spencer: Error =

7/17

Derandomizing Chernoff(Pessimistic estimators, exp. moments, hyp. cosine rule, …)

•  

8/17

The Problem

9/17

Such approaches cannot get rid of

(Chooser-Pusher Games: Where each column rounded in an online manner)

Algorithm of Bansal uses a more global approach

10/17

Algorithm (at high level)

Cube: {-1,1}nstart

finish

Each dimension: A variableEach vertex: A rounding

Algorithm: At step t, update Fix variable if reaches -1 or 1.

g: random is random Gaussian in

Each distributed as a GaussianBut the ’s are correlated.

11/17

SDP relaxations

SDPs (LP on )“is small” 8 j |vi|2 = 1 Intended soln. vi = (+1,0,…,0) or (-1,0,…,0).

Spencer’s result (entropy method) guarantees feasibility.

Key point of Gaussian rounding: Say if Then

12/17

Analysis (at high level)

Cube: {-1,1}n

Analysis: Progress: Few steps to reach a vertex (walk has high variance)

Low Discrepancy: For each equation, discrepancy random walk has low variance

start

finish

Each dimension: An ElementEach vertex: A Coloring

Making it Deterministic

Need to find an update thati) Makes Progress ii) Adds low discrepancy to equations.

Recall, for Chernoff: Round one variable at a time (Progress)

Whether -1 or +1, guided by the potential. (Low Discrepancy)

13/17

Tracking the propertiesi) For low discrepancy. Define suitable potential and bound its increase (as in Chernoff, but refined)

ii) For Progress Potential Energy should go up sufficiently

Conflicting goals (hold in expectation)No reason why such an update should even exist.

14/17

Our fix

Add extra constraints to SDP to force a good update to exist.

Orthogonality trick: Say currently at Add SDP constraint

Ensures that update orthogonal to x. The length (potential) always increases!

Analogous constraint for low discrepancy potential(bounds increase by right amount)

15/17

origin

x(t-1)

x(t): Newposition

𝛿 (𝑡 )

Trouble

Why should this SDP remain feasible?

In Bansal’s (randomized) algorithm SDP was feasible due to Spencer’s existential result.

Key point: New constraint is of similar type i.e.is small)Entropy method: new SDP is still feasible.

16/17

Finish off: Use k-wise vectors instead of Gaussian

17/17

Concluding Remarks

Idea: Add new constraints to force a deterministic choice to exist.

Works more generally for other discrepancy problems.

Can potentially have other applications.

Thank You!

Techniques

18/17

Entropy Method

Spencer’s Result

Bansal’s Result

SDPs

New “orthogonality” idea (based on entropy)+ K-wise independence, pessimistic estimators, …

This Result