Estimating
Distance to Monotonicity
Seshadhri Comandur, Princeton University
In standard property testing, the task is to distinguish
between
objects that have a property P and those that are \eps-far from P, for
some
\eps > 0. In this setting, it is perfectly acceptable for the tester
to
provide a negative answer for every input object that does not satisfy
P.
This implies that property testing in and of itself cannot be expected
to
yield any information whatsoever about the distance from the object to
the
property. We address this problem in this paper, restricting our
attention
to monotonicity testing. A function f:[1,n] -> R is at distance
\eps_f from
being monotone if it can (and must) be modified at \eps_f*n places to
become
monotone. For any fixed \delta>0, we compute, with probability at
least 2/3,
an interval [(1/2 -\delta)\eps, \eps] that encloses \eps_f. The running
time
of our algorithm is linear in log n and almost linear in 1/eps_f, which
is a
substantial improvement over previous work.
Joint with with Nir Ailon, Bernard Chazelle and Ding Liu
|
Recent
progress on algorithms for multicommodity routing problems
Chandra Chekuri, Lucent Bell Labs.
Abstract:
A fundamental problem in combinatorial optimization is
the edge-disjoint paths problem. We are given a graph G=(V,E)
and set of pairs of vertices (s_1,t_1), (s_2,t_2), ..., (s_k,t_k).
The objective is to decide if all the pairs can be connected
by edge-disjoint paths. Generalizations include the unsplittable
flow problem and the maximum integer multicommodity flow problem.
All of the above problems are NP-hard and we focus on approximation
algorithms for the maximization versions where the goal is to
maximize the number of pairs that can be connected (routed).
In this talk we give an overview of some recent results and
techniques that give substantially improved results for these
problems in undirected planar graphs if two paths are allowed
on an edge. Along the way we introduce a previously unexplored
variant called the all-or-nothing multicommodity flow problem
and discuss results for this problem.
Joint work with Sanjeev Khanna (U. Penn) and Bruce Shepherd (Bell
Labs).
|
Euclidean
Distortion and Sparsest Cut
James Lee, UC Berkeley
Embeddings of finite metric spaces into geometric spaces is a
burgeoning
research area with many applications to algorithm design and pure
mathematics. Two of the fundamental problems in this field concern
embedding metrics of negative type into L_1, and embedding finite
subsets of L_1 into a Euclidean space. The former problem has
come to
prominence because of its intimate relationship with computing the
Sparsest Cut in graphs, and the latter because of its role in the local
theory of Banach spaces. In this talk, I will present recent
progress on
both questions. In particular, I will discuss our recent proof
that every
n-point metric of negative type has Euclidean distortion which is
bounded
above by sqrt{log n} log log n, a result which is optimal up to the
log log n factor. This leads immediately to an O(sqrt{log n} log
log n)-
approximation for the Sparsest Cut problem with general demands.
[This is joint work with Sanjeev Arora and Assaf Naor.]
---------------- |
Sampling to estimate arbitrary subset sums
Mikkel Thorup
(joint work with Nick Duffield and Carsten Lund)
Abstract
Starting with a set of weighted items, we want to create a generic sample of
a certain size that we can later use to estimate the total weight of
arbitrary subsets. Applied in Internet traffic analysis, the items
could be records summarizing the flows streaming by a router, with,
say, a hundred records sampled each hour. A subset could be flow
records of a worm attack identified later. Our past samples now
allow us to trace the history of the attack even though the worm was
unknown while the samples were made.
Estimation from the samples must be accurate even with heavy-tailed
distributions where most of the weight is concentrated on a few
heavy items. We want the sample to be weight sensitive, giving
priority to heavy items. At the same time, we want sampling without
replacement in order to avoid selecting heavy items multiple times.
To fulfill these requirements we introduce priority sampling, which
is the first weight sensitive sampling scheme without replacement
that is suitable for estimating subset sums. Testing priority
sampling on Internet traffic analysis, we found it to perform orders
of magnitude better than previous schemes.
Priority sampling is simple to define and implement: we consider a steam of
items i=0,...,n-1 with weights w_i. For each item i, we generate
a random number r_i in (0,1) and create a priority q_i=w_i/r_i.
The sample S consists of the k highest priority items. Let
t be the (k+1)th highest priority. Each sampled item i in S
gets a weight estimate W_i=max{w_i,t}, while non-sampled
items get weight estimate W_i=0.
Magically, it turns out that the weight estimates are unbiased, that
is, E[W_i]=w_i, and by linearity of expectation, we get unbiased
estimators over any subset sum simply by adding the sampled weight
estimates from the subset. Also, we can estimate the variance of the
estimates, and surpricingly, there is no co-variance between different
weight estimates W_i and W_j.
We conjecture an extremely strong near-optimality; namely that for
any weight sequence, there exists no specialized scheme for
sampling k items with unbiased estimators that gets smaller
total variance than priority sampling with k+1 items.
|
O(\sqrt{log
n}) approximation algorithms for Min UnCut, Min 2CNF Deletion,
and directed cut problems
Yury Makarychev
joint work with Amit Agarwal, Moses Charikar, Konstantin Makarychev
Abstract:
We give $O(\sqrt{\log n})$-approximation algorithms for the Min UnCut,
Min 2CNF Deletion, Directed Balanced Separator, and Directed Sparsest
Cut problems. The previously best known algorithms give an $O(\log
n)$-approximation
for Min UnCut, Directed Balanced Separator, Directed Sparsest Cut,
and an $O(\log n \log\log n)$-approximation for Min 2CNF Deletion.
We also show that the integrality gap of an SDP relaxation of the
Minimum Multicut problem is $\Omega(\log n)$. |
Experts in a Markov Decision Process
We consider an MDP setting in which the reward function is allowed to
change during each time step of play (possibly in an adversarial
manner), yet the dynamics remain fixed. Similar to the experts
setting, we address the question of how well can an agent do when
compared to the reward achieved under the best stationary policy over
time. We provide \emph{efficient} algorithms, which have regret bounds
with \emph{no dependence} on the size of state space. Instead, these
bounds depend only on a certain horizon time of the process and
logarithmically on the number of actions. We also show that in the
case that the dynamics change over time, the problem becomes
computationally hard
Joint work with Sham Kakade and Yishay Mansour
|
Hardness
of Three Routing Problems
Matt Andrews, Lucent Bell Labs.
In this talk we present hardness-of-approximation results for the
undirected
edge-disjoint paths problem, the undirected congestion minimization
problem
and the buy-at-bulk network design problem. We first show that the
problems
are hard to approximate when each demand is constrained to follow one
of a
small set of "canonical paths". We then show how to embed our
examples into
high-girth graphs in such a way that all good routings use many
canonical
paths. These techniques allow us to derive hardness results for the
general
problems.
|
Eli Berger, IAS.
Title:
Menger's
theorem for infinite graphs
Abstract:
We prove an old conjecture of Erdos, saying that Menger's
theorem is valid also for infinite graphs, in the following strong form:
given sets A and B of vertices in a graph (possibly directed,
possibly infinite), there exists a family P of disjoint A-B
paths, and an A-B separating set S, such that S consists
of a choice of precisely one vertex from each path in P.
The talk will describe the history of the problem and the main ideas of
our proof.
This is joint work with Ron Aharoni
|
The
Postselection Principle
Scott Aaronson (IAS)
If we want to learn an unknown object, then we should repeatedly
measure the object in such a way that we gain information, even
conditioned on all of the previous measurements. I'll show how
this simple, obvious principle underlies:
* An iterative learning algorithm of Bshouty et al., which Kobler and
Watanabe used to improve the Karp-Lipton Theorem and to show that for
all k, the class ZPP^NP does not have circuits of size n^k.
* A recent improvement of mine, to ZPP with parallel NP queries.
* A result of mine that BQP/qpoly is contained in PP/poly: that is,
quantum computers with polynomial-size quantum advice can be simulated
in PP with polynomial-size classical advice.
* Quantum circuit lower bounds and a quantum Karp-Lipton theorem.
|
Active Learning â Theory and Practice
Ran Gilad-Bachrach, The Hebrew University
Abstract:
Passive learners only listen to their teachers whereas active learners can direct questions to them. Both theory and practice confirm that active learners can significantly outperform their passive contenders. In this talk, we will review the latest results in this field. We will look at different models for active learning and discuss their pros and cons. We will show that when certain conditions apply, active learners learn exponentially faster than passive learners do. We will focus on our attempt to come up with an algorithm that has both theoretical grounding and efficient implementation.
The talk will be self-contained.
|
Holographic
Algorithms
Leslie G. Valiant
Harvard University
Using the notion of polynomial time reduction
computer
scientists have discovered an astonishingly rich web of
interrelationships
among the myriad natural computational problems that arise in diverse
applications. These relationships have been used both to give evidence
of
intractability, such as that of NP-completeness, as well as some
surprising new
algorithms.
In this talk we discuss a
notion of reduction, which we call
a holographic reduction, that is more general than the traditional one.
Instead
of locally mapping solutions one-to-one it maps them many-to-many but
preserves
the sum of the solutions. One application is to finding new polynomial time algorithms where none was known before. We shall give some examples
of such algorithms.
A more radical potential
direction is that of revisiting the
currently accepted conjectures of
computer science, such as that P does not equal NP, and seeing whether
this new
kind of reduction offers any new insights towards either positive or
negative
results. The talk will review complexity theory in this light.
|
On Embeddability of
Negative Type Metrics into L_1
-----------------------------------------------------------
Subhash Khot, Georgia Tech
Goemans and Linial conjectured that negative type metrics
(metrics that are squares of Euclidean metrics) embed into L_1
with constant distortion. Negative type metrics arise naturally
as solutions of semidefinite relaxations for Sparsest Cut and
Graph Partitioning problems. The GL-conjecture implies that the
"integrality ratio" for the SDP-relaxation is bounded by a
constant (which gives constant factor approximation algorithm).
The recent breakthrough algorithm of [Arora Rao Vazirani] gave
O(\sqrt{log n}) upper bound on the integrality ratio and they
too conjectured a constant upper bound.
We disprove both the above conjectures by constructing an
integrality ratio example with ratio (log log n)^c for some
constant c > 0 (the disproof of ARV conjectrue holds only for the
so-called "non-uniform" version of Sparsest Cut). Surprisingly,
our construction is inspired from complexity theoretic tools,
namely PCPs, Fourier Analysis, and the Unique Games Conjecture
(UGC) by [Khot].
In this talk, I will give a overview of our construction
and other research that UGC has led to (e.g. optimal hardness
results for Vertex Cover and MAX-CUT assuming UGC). The talk will
be self-contained.
Mostly based on joint work with Nisheeth Vishnoi.
|
On Non-uniform
Multicommodity Buy-at-Bulk Network Design
Adriana Karagiozova, Princeton University
We study the multicommodity buy-at-bulk network design problem
where the goal is to buy capacity on edges of a network so as
to enable the demands between a given set of source-sink pairs
to be routed - the objective is to minimize the cost of such a
solution. The key aspect of this problem is that the cost of an
edge in the graph is a concave monotone function of the flow
cross the edge and hence exhibits economy of scale - it pays
to aggregate flow paths as much as possible. In the non-uniform
case, each edge has its own cost function, possibly different
from other edges.
Special cases of this problem have been studied extensively.
We present the first non-trivial approximation algorithm for the
general case. Our algorithm is an extremely simple randomized
greedy algorithm, involving not much more than shortest path
computations. We achieve an approximation guarantee of
exp(O(sqrt(log(n)log(log(n)))) where n is the total demand in
the graph.
This is joint work with Moses Charikar |
The Complexity of
Online Memory Checking
Guy Rothblum
Suppose you want to store a large file on a remote and unreliable
server. You would like to verify that your file has not been
corrupted, so you store a small private (randomized)``fingerprint'' of
the file on your own computer. This is the setting for the
well-studied authentication problem, and the size of the required
private fingerprint is well understood. We study the problem of
sub-linear authentication: suppose you would like to encode and store
your file in a way that allows you to verify that it has not been
corrupted, but without reading all of it. If you only want to read t
bits of the file, how large does the size s of the fingerprint need to
be? We define this problem formally, and show a tight lower bound on
the relationship between s and t when the adversary is not
computationally bounded, namely: s x t= Omega(n) where n is the file
size. This is an easier case of the online memory checking problem,
introduced by Blum, Evans, Gemmel, Kannan and Naor in 1991, and hence
the same (tight) lower bound applies also to this problem.
It was previously shown that when the adversary is not computationally
bounded, under the assumption that one-way functions exist, it is
possible to construct much better online memory checkers and
sub-linear authentication schemes. We show that the existence of
one-way functions is also a necessary condition: even slightly
breaking the s x t= Omega(n) lower bound in a computational
setting
implies the existence of one-way functions.
(Joint work with Moni Naor) |
Building Cayley expanders for arbitrary
groups using randomness- efficient sampling
David Xiao, Princeton
Suppose we are given an arbitrary group G of size n and we are
asked to construct a Cayley graph with good expansion, i.e.
second-largest eigenvalue bounded by a constant away from
1. Take O(log n) random elements from G and call this set
S. Alon and Roichman showed that with high probability, the
Cayley graph given by G using S as the generating set is a good
expander.
We derandomize this to give a deterministic polynomial time
algorithm that for an arbitrary group G finds a set S of size
O(log n) such that the Cayley graph of G using S is a good
expander. Previously it was only known how to find S of
size n^c for arbitrary c > 0, due to Shpilka and Wigderson.
Our technique is to use the expander walk sampler of Ajtai,
Komlos, and Szemeredi. Our analysis shows that this sampler
is also good at sampling matrix-valued functions, and not just
the boolean-valued functions considered by AKS or the real-valued
functions considered by Gillman. The expander walk sampler
reduces the randomness complexity enough to allow us to enumerate
over all walks and find a suitable S in polynomial time.
Finally we discuss possible other applications of this new use of
the AKS sampler.
This is joint work with Avi Wigderson.
|
|
|
|
|
|
|
|
|
|