Princeton Seminars in Theoretical Computer Science, 
Fall 2004

(All events held at the Computer Science Department, 35 Olden St., unless stated otherwise)
To subscribe to the mailing list for seminar announcements, visit this page

 
              Schedule 
Sept 17 Seshadhri Commandur, Princeton
Estimating Distance to Monotonicity
Oct 1 Satyen Kale and Elad Hazan, Princeton
The multiplicative update rule: a meta algorithm.
Oct 8 Chandra Chekuri, Lucent Bell Labs.
Recent progress on algorithms for multicommodity routing problems
Oct 22 James Lee, UC Berkeley
Euclidean Distortion and Sparsest Cut
Oct 29 No theory lunch (Fall break)
Nov 5 No theory lunch; informal theory get together at prospect house
Nov 12 Sampling to estimate arbitrary subset sums
Mikkel Thorup, AT &T
Nov 19 No theory lunch (Theory Day at Columbia U.)
Nov 25 No theory lunch; Thanksgiving
Dec 3 O(\sqrt{log n}) approximation algorithms for Min UnCut, Min 2CNF  Deletion, and directed cut problems. Yury Makarychev
Dec 10
Sham Kakade (University of Pennsylvania)
Deterministic Calibration and Nash Equilibrium
February 4
Eyal Even-Dar, Tel Aviv
Experts in a Markov Decision Process.

Feb 11 Matthew Andrews, Lucent Bell Labs.
Hardness of Three Routing Problems
Feb 18
Eli Berger, IAS. 
Menger's theorem for infinite graphs.
Feb25
Scott Aaronson, IAS
The Postselection Principle
March 4
Ran Gilad-Bachrach, Hebrew University
Active Learning :Theory and Practice
March 11 Leslie Valiant, Harvard.   Holographic Algorithms
March 18 No lunch (Spring break)
March 25 Subhash Khot, Georgia Tech
On embeddability of negative type metrics into l_1
April 22
On Non-uniform Multicommodity Buy-at-Bulk Network Design

Adriana Karagiozova, Princeton University
April 29
Guy Rothblum, Weizmann Institute
Complexity of online memory checking
May 6
Elad Hazan
May 27
Luca Trevisan
Uniform amplification of hardness in NP
June 17
David Xiao
Building Cayley expanders for arbitrary groups using randomness- efficient sampling


 

 
           Abstracts
Estimating Distance to Monotonicity

Seshadhri Comandur, Princeton University

In standard property testing, the task is to distinguish between
objects that have a property P and those that are \eps-far from P, for some
\eps > 0. In this setting, it is perfectly acceptable for the tester to
provide a negative answer for every input object that does not satisfy P.
This implies that property testing in and of itself cannot be expected to
yield any information whatsoever about the distance from the object to the
property. We address this problem in this paper, restricting our attention
to monotonicity testing. A function f:[1,n] -> R is at distance \eps_f from
being monotone if it can (and must) be modified at \eps_f*n places to become
monotone. For any fixed \delta>0, we compute, with probability at least 2/3,
an interval [(1/2 -\delta)\eps, \eps] that encloses \eps_f. The running time
of our algorithm is linear in log n and almost linear in 1/eps_f, which is a
substantial improvement over previous work.

Joint with with Nir Ailon, Bernard Chazelle and Ding Liu

 

  Recent progress on algorithms for multicommodity routing problems
 Chandra Chekuri, Lucent Bell Labs.


Abstract:
A fundamental problem in combinatorial optimization is
the edge-disjoint paths problem. We are given a graph G=(V,E)
and set of pairs of vertices (s_1,t_1), (s_2,t_2), ..., (s_k,t_k).
The objective is to decide if all the pairs can be connected
by edge-disjoint paths. Generalizations include the unsplittable
flow problem and the maximum integer multicommodity flow problem.

All of the above problems are NP-hard and we focus on approximation
algorithms for the maximization versions where the goal is to
maximize the number of pairs that can be connected (routed).

In this talk we give an overview of some recent results and
techniques that give substantially improved results for these
problems in undirected planar graphs if two paths are allowed
on an edge. Along the way we introduce a previously unexplored
variant called the all-or-nothing multicommodity flow problem
and discuss results for this problem.


Joint work with Sanjeev Khanna (U. Penn) and Bruce Shepherd (Bell Labs).
 

 

Euclidean Distortion and Sparsest Cut

James Lee, UC Berkeley

Embeddings of finite metric spaces into geometric spaces is a burgeoning
research area with many applications to algorithm design and pure mathematics. Two of the fundamental problems in this field concern embedding metrics of negative type into L_1, and embedding finite
subsets of L_1 into a Euclidean space.  The former problem has come to
prominence because of its intimate relationship with computing the
Sparsest Cut in graphs, and the latter because of its role in the local
theory of Banach spaces.  In this talk, I will present recent progress on
both questions.  In particular, I will discuss our recent proof that every
n-point metric of negative type has Euclidean distortion which is bounded
above by sqrt{log n} log log n, a result which is optimal up to the
log log n factor.  This leads immediately to an O(sqrt{log n} log log n)-
approximation for the Sparsest Cut problem with general demands.

[This is joint work with Sanjeev Arora and Assaf Naor.]
----------------
Sampling to estimate arbitrary subset sums
Mikkel Thorup (joint work with Nick Duffield and Carsten Lund) Abstract Starting with a set of weighted items, we want to create a generic sample of a certain size that we can later use to estimate the total weight of arbitrary subsets. Applied in Internet traffic analysis, the items could be records summarizing the flows streaming by a router, with, say, a hundred records sampled each hour. A subset could be flow records of a worm attack identified later. Our past samples now allow us to trace the history of the attack even though the worm was unknown while the samples were made. Estimation from the samples must be accurate even with heavy-tailed distributions where most of the weight is concentrated on a few heavy items. We want the sample to be weight sensitive, giving priority to heavy items. At the same time, we want sampling without replacement in order to avoid selecting heavy items multiple times. To fulfill these requirements we introduce priority sampling, which is the first weight sensitive sampling scheme without replacement that is suitable for estimating subset sums. Testing priority sampling on Internet traffic analysis, we found it to perform orders of magnitude better than previous schemes. Priority sampling is simple to define and implement: we consider a steam of items i=0,...,n-1 with weights w_i. For each item i, we generate a random number r_i in (0,1) and create a priority q_i=w_i/r_i. The sample S consists of the k highest priority items. Let t be the (k+1)th highest priority. Each sampled item i in S gets a weight estimate W_i=max{w_i,t}, while non-sampled items get weight estimate W_i=0. Magically, it turns out that the weight estimates are unbiased, that is, E[W_i]=w_i, and by linearity of expectation, we get unbiased estimators over any subset sum simply by adding the sampled weight estimates from the subset. Also, we can estimate the variance of the estimates, and surpricingly, there is no co-variance between different weight estimates W_i and W_j. We conjecture an extremely strong near-optimality; namely that for any weight sequence, there exists no specialized scheme for sampling k items with unbiased estimators that gets smaller total variance than priority sampling with k+1 items.

 

O(\sqrt{log n}) approximation algorithms for Min UnCut, Min 2CNF  Deletion, and directed cut problems

Yury Makarychev
joint work with Amit Agarwal, Moses Charikar, Konstantin Makarychev

Abstract:
We give $O(\sqrt{\log n})$-approximation algorithms for the Min UnCut,
Min 2CNF Deletion, Directed Balanced Separator, and Directed Sparsest Cut problems. The previously best known algorithms give an $O(\log n)$-approximation
for Min UnCut, Directed Balanced Separator, Directed Sparsest Cut,
and an $O(\log n \log\log n)$-approximation for Min 2CNF Deletion.

We also show that the integrality gap of an SDP relaxation of the Minimum Multicut problem is $\Omega(\log n)$.
Experts in a Markov Decision Process
We consider an MDP setting in which the reward function is allowed to change during each time step of play (possibly in an adversarial manner), yet the dynamics remain fixed. Similar to the experts setting, we address the question of how well can an agent do when compared to the reward achieved under the best stationary policy over time. We provide \emph{efficient} algorithms, which have regret bounds with \emph{no dependence} on the size of state space. Instead, these bounds depend only on a certain horizon time of the process and logarithmically on the number of actions. We also show that in the case that the dynamics change over time, the problem becomes computationally hard Joint work with Sham Kakade and Yishay Mansour
 

  Hardness of Three Routing Problems
 Matt Andrews, Lucent Bell Labs.


In this talk we present hardness-of-approximation results for the undirected
edge-disjoint paths problem, the undirected congestion minimization problem
and the buy-at-bulk network design problem. We first show that the problems
are hard to approximate when each demand is constrained to follow one of a
small set of "canonical paths".  We then show how to embed our examples into
high-girth graphs in such a way that all good routings use many canonical
paths. These techniques allow us to derive hardness results for the general
problems.

Eli Berger, IAS. 

Title:
Menger's theorem for infinite graphs

Abstract:

We prove an old conjecture of Erdos, saying that Menger's
theorem is valid also for infinite graphs, in the following strong form:
given sets A and B of vertices in a graph (possibly directed, possibly  infinite), there exists a family P of disjoint A-B paths, and an A-B separating set S, such that S consists
of a choice of precisely one vertex from each path in P.
The talk will describe the history of the problem and the main ideas of
our proof.

This is joint work with Ron Aharoni

The Postselection Principle
Scott Aaronson (IAS)

If we want to learn an unknown object, then we should repeatedly measure the object in such a way that we gain information, even conditioned on all of the previous measurements.  I'll show how this simple, obvious principle underlies:

* An iterative learning algorithm of Bshouty et al., which Kobler and Watanabe used to improve the Karp-Lipton Theorem and to show that for all k, the class ZPP^NP does not have circuits of size n^k.

* A recent improvement of mine, to ZPP with parallel NP queries.

* A result of mine that BQP/qpoly is contained in PP/poly: that is, quantum computers with polynomial-size quantum advice can be simulated in PP with polynomial-size classical advice.

* Quantum circuit lower bounds and a quantum Karp-Lipton theorem.
Active Learning â Theory and Practice

Ran Gilad-Bachrach, The Hebrew University

Abstract:

Passive learners only listen to their teachers whereas active learners
can direct questions to them. Both theory and practice confirm that
active learners can significantly outperform their passive contenders.
In this talk, we will review the latest results in this field. We will
look at different models for active learning and discuss their pros
and cons. We will show that when certain conditions apply, active
learners learn exponentially faster than passive learners do. We will
focus on our attempt to come up with an algorithm that has both
theoretical grounding and efficient implementation.

The talk will be self-contained.

Holographic Algorithms

Leslie G. Valiant

Harvard University

Using the notion of polynomial time reduction computer scientists have discovered an astonishingly rich web of interrelationships among the myriad natural computational problems that arise in diverse applications. These relationships have been used both to give evidence of intractability, such as that of NP-completeness, as well as some surprising new algorithms.

In this talk we discuss a notion of reduction, which we call a holographic reduction, that is more general than the traditional one. Instead of locally mapping solutions one-to-one it maps them many-to-many but preserves the sum of the solutions. One application is to finding new  polynomial time algorithms where none  was known before. We shall give some examples of such algorithms.

A more radical potential direction is that of revisiting the currently accepted conjectures  of computer science, such as that P does not equal NP, and seeing whether this new kind of reduction offers any new insights towards either positive or negative results. The talk will review complexity theory in this light.

 

On Embeddability of Negative Type Metrics into L_1
-----------------------------------------------------------
Subhash Khot, Georgia Tech

Goemans and Linial conjectured that negative type metrics
(metrics that are squares of Euclidean metrics) embed into L_1
with constant distortion. Negative type metrics arise naturally
as solutions of semidefinite relaxations for Sparsest Cut and
Graph Partitioning problems. The GL-conjecture implies that the
"integrality ratio" for the SDP-relaxation is bounded by a
constant (which gives constant factor approximation algorithm).
The recent breakthrough algorithm of [Arora Rao Vazirani] gave
O(\sqrt{log n}) upper bound on the integrality ratio and they
too conjectured a constant upper bound.

We disprove both the above conjectures by constructing an
integrality ratio example with ratio (log log n)^c for some
constant c > 0 (the disproof of ARV conjectrue holds only for the
so-called "non-uniform" version of Sparsest Cut). Surprisingly,
our construction is inspired from complexity theoretic tools,
namely PCPs, Fourier Analysis, and the Unique Games Conjecture
(UGC) by [Khot].

In this talk, I will give a overview of our construction
and other research that UGC has led to (e.g. optimal hardness
results for Vertex Cover and MAX-CUT assuming UGC). The talk will
be self-contained.

Mostly based on joint work with Nisheeth Vishnoi.

On Non-uniform Multicommodity Buy-at-Bulk Network Design

Adriana Karagiozova, Princeton University

We study the multicommodity buy-at-bulk network design problem
where the goal is to buy capacity on edges of a network so as
to enable the demands between a given set of source-sink pairs
to be routed - the objective is to minimize the cost of such a
solution. The key aspect of this problem is that the cost of an
edge in the graph is a concave monotone function of the flow
cross the edge and hence exhibits economy of scale - it pays
to aggregate flow paths as much as possible. In the non-uniform
case, each edge has its own cost function, possibly different
from other edges.

Special cases of this problem have been studied extensively.
We present the first non-trivial approximation algorithm for the
general case. Our algorithm is an extremely simple randomized
greedy algorithm, involving not much more than shortest path
computations. We achieve an approximation guarantee of
exp(O(sqrt(log(n)log(log(n)))) where n is the total demand in
the graph.

This is joint work with Moses Charikar
The Complexity of Online Memory Checking
Guy Rothblum

Suppose you want to store a large file on a remote and unreliable
server. You would like to verify that your file has not been
corrupted, so you store a small private (randomized)``fingerprint'' of
the file on your own computer. This is the setting for the
well-studied authentication problem, and the size of the required
private fingerprint is well understood. We study the problem of
sub-linear authentication: suppose you would like to encode and store
your file in a way that allows you to verify that it has not been
corrupted, but without reading all of it. If you only want to read t
bits of the file, how large does the size s of the fingerprint need to
be? We define this problem formally, and show a tight lower bound on
the relationship between s and t when the adversary is not
computationally bounded, namely: s x t= Omega(n) where n is the file
size. This is an easier case of the online memory checking problem,
introduced by Blum, Evans, Gemmel, Kannan and Naor in 1991, and hence
the same (tight) lower bound applies also to this problem.

It was previously shown that when the adversary is not computationally
bounded, under the assumption that one-way functions exist, it is
possible to construct much better online memory checkers and
sub-linear authentication schemes. We show that the existence of
one-way functions is also a necessary condition: even slightly
breaking the  s x t= Omega(n) lower bound in a computational setting
implies  the existence of one-way functions.

(Joint work with Moni Naor)
Building Cayley expanders for arbitrary groups using randomness- efficient sampling

David Xiao, Princeton

Suppose we are given an arbitrary group G of size n and we are asked  to construct a Cayley graph with good expansion, i.e. second-largest  eigenvalue bounded by a constant away from 1.  Take O(log n) random  elements from G and call this set S.  Alon and Roichman showed that  with high probability, the Cayley graph given by G using S as the  generating set is a good expander.

We derandomize this to give a deterministic polynomial time algorithm  that for an arbitrary group G finds a set S of size O(log n) such  that the Cayley graph of G using S is a good expander.  Previously it  was only known how to find S of size n^c for arbitrary c > 0, due to  Shpilka and Wigderson.

Our technique is to use the expander walk sampler of Ajtai, Komlos,  and Szemeredi.  Our analysis shows that this sampler is also good at  sampling matrix-valued functions, and not just the boolean-valued  functions considered by AKS or the real-valued functions considered  by Gillman.  The expander walk sampler reduces the randomness  complexity enough to allow us to enumerate over all walks and find a  suitable S in polynomial time.

Finally we discuss possible other applications of this new use of the  AKS sampler.

This is joint work with Avi Wigderson.

 


 


 

 

 

 

 

 
 (archive of past announcements)