Results 1  10
of
190
A Critical Point For Random Graphs With A Given Degree Sequence
, 2000
"... Given a sequence of nonnegative real numbers 0 ; 1 ; : : : which sum to 1, we consider random graphs having approximately i n vertices of degree i. Essentially, we show that if P i(i \Gamma 2) i ? 0 then such graphs almost surely have a giant component, while if P i(i \Gamma 2) i ! 0 the ..."
Abstract

Cited by 289 (7 self)
 Add to MetaCart
Given a sequence of nonnegative real numbers 0 ; 1 ; : : : which sum to 1, we consider random graphs having approximately i n vertices of degree i. Essentially, we show that if P i(i \Gamma 2) i ? 0 then such graphs almost surely have a giant component, while if P i(i \Gamma 2) i ! 0 then almost surely all components in such graphs are small. We can apply these results to G n;p ; G n;M , and other wellknown models of random graphs. There are also applications related to the chromatic number of sparse random graphs.
The Power of Two Choices in Randomized Load Balancing
 IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS
, 1996
"... Suppose that n balls are placed into n bins, each ball being placed into a bin chosen independently and uniformly at random. Then, with high probability, the maximum load in any bin is approximately log n log log n . Suppose instead that each ball is placed sequentially into the least full of d ..."
Abstract

Cited by 201 (23 self)
 Add to MetaCart
Suppose that n balls are placed into n bins, each ball being placed into a bin chosen independently and uniformly at random. Then, with high probability, the maximum load in any bin is approximately log n log log n . Suppose instead that each ball is placed sequentially into the least full of d bins chosen independently and uniformly at random. It has recently been shown that the maximum load is then only log log n log d +O(1) with high probability. Thus giving each ball two choices instead of just one leads to an exponential improvement in the maximum load. This result demonstrates the power of two choices, and it has several applications to load balancing in distributed systems. In this thesis, we expand upon this result by examining related models and by developing techniques for stu...
On the Generalization Ability of Online Learning Algorithms
 IEEE Transactions on Information Theory
, 2001
"... In this paper we show that online algorithms for classification and regression can be naturally used to obtain hypotheses with good datadependent tail bounds on their risk. Our results are proven without requiring complicated concentrationofmeasure arguments and they hold for arbitrary onlin ..."
Abstract

Cited by 133 (8 self)
 Add to MetaCart
In this paper we show that online algorithms for classification and regression can be naturally used to obtain hypotheses with good datadependent tail bounds on their risk. Our results are proven without requiring complicated concentrationofmeasure arguments and they hold for arbitrary online learning algorithms. Furthermore, when applied to concrete online algorithms, our results yield tail bounds that in many cases are comparable or better than the best known bounds.
The Size Of The Giant Component Of A Random Graph With A Given Degree Sequence
 COMBIN. PROBAB. COMPUT
, 2000
"... Given a sequence of nonnegative real numbers 0 ; 1 ; : : : which sum to 1, we consider a random graph having approximately i n vertices of degree i. In [12] the authors essentially show that if P i(i \Gamma 2) i ? 0 then the graph a.s. has a giant component, while if P i(i \Gamma 2) i ! 0 ..."
Abstract

Cited by 127 (0 self)
 Add to MetaCart
Given a sequence of nonnegative real numbers 0 ; 1 ; : : : which sum to 1, we consider a random graph having approximately i n vertices of degree i. In [12] the authors essentially show that if P i(i \Gamma 2) i ? 0 then the graph a.s. has a giant component, while if P i(i \Gamma 2) i ! 0 then a.s. all components in the graph are small. In this paper we analyze the size of the giant component in the former case, and the structure of the graph formed by deleting that component. We determine
Using Confidence Bounds for ExploitationExploration Tradeoffs
 Journal of Machine Learning Research
, 2002
"... We show how a standard tool from statistics  namely confidence bounds  can be used to elegantly deal with situations which exhibit an exploitationexploration tradeo#. Our technique for designing and analyzing algorithms for such situations is general and can be applied when an algorithm h ..."
Abstract

Cited by 109 (2 self)
 Add to MetaCart
We show how a standard tool from statistics  namely confidence bounds  can be used to elegantly deal with situations which exhibit an exploitationexploration tradeo#. Our technique for designing and analyzing algorithms for such situations is general and can be applied when an algorithm has to make exploitationversusexploration decisions based on uncertain information provided by a random process.
Building LowDiameter P2P Networks
, 2001
"... In a peertopeer (P2P) network, nodes connect into an existing network and participate in providing and availing of services. There is no dichotomy between a central server and distributed clients. Current P2P networks (e.g., Gnutella) are constructed by participants following their own uncoordina ..."
Abstract

Cited by 108 (2 self)
 Add to MetaCart
In a peertopeer (P2P) network, nodes connect into an existing network and participate in providing and availing of services. There is no dichotomy between a central server and distributed clients. Current P2P networks (e.g., Gnutella) are constructed by participants following their own uncoordinated (and often whimsical) protocols; they consequently suffer from frequent network overload and fragmentation into disconnected pieces separated by chokepoints with inadequate bandwidth. In this paper we propose a simple scheme for participants to build P2P networks in a distributed fashion, and prove that it results in connected networks of constant degree and logarithmic diameter. It does so with no global knowledge of all the nodes in the network. In the most common P2P application to date (search), these properties are crucial.
Resilient Multicast using Overlays
 In Proc. of ACM Sigmetrics
, 2003
"... (PRM): a multicast data recovery scheme that improves data delivery ratios while maintaining low endtoend latencies. PRM has both a proactive and a reactive components; in this paper we describe how PRM can be used to improve the performance of applicationlayer multicast protocols especially when ..."
Abstract

Cited by 106 (8 self)
 Add to MetaCart
(PRM): a multicast data recovery scheme that improves data delivery ratios while maintaining low endtoend latencies. PRM has both a proactive and a reactive components; in this paper we describe how PRM can be used to improve the performance of applicationlayer multicast protocols especially when there are high packet losses and host failures. Through detailed analysis in this paper, we show that this loss recovery technique has efficient scaling properties—the overheads at each overlay node asymptotically decrease to zero with increasing group sizes. As a detailed case study, we show how PRM can be applied to the NICE applicationlayer multicast protocol. We present detailed simulations of the PRMenhanced NICE protocol for 10 000 node Internetlike topologies. Simulations show that PRM achieves a high delivery ratio ( 97%) with a low latency bound (600 ms) for environments with high endtoend network losses (1%–5%) and high topology change rates (5 changes per second) while incurring very low overheads ( 5%). Index Terms—Multicast, networks, overlays, probabilistic forwarding, protocols, resilience. I.
Stochastic Approximation Approach to Stochastic Programming
"... In this paper we consider optimization problems where the objective function is given in a form of the expectation. A basic difficulty of solving such stochastic optimization problems is that the involved multidimensional integrals (expectations) cannot be computed with high accuracy. The aim of th ..."
Abstract

Cited by 99 (14 self)
 Add to MetaCart
In this paper we consider optimization problems where the objective function is given in a form of the expectation. A basic difficulty of solving such stochastic optimization problems is that the involved multidimensional integrals (expectations) cannot be computed with high accuracy. The aim of this paper is to compare two computational approaches based on Monte Carlo sampling techniques, namely, the Stochastic Approximation (SA) and the Sample Average Approximation (SAA) methods. Both approaches, the SA and SAA methods, have a long history. Current opinion is that the SAA method can efficiently use a specific (say linear) structure of the considered problem, while the SA approach is a crude subgradient method which often performs poorly in practice. We intend to demonstrate that a properly modified SA approach can be competitive and even significantly outperform the SAA method for a certain class of convex stochastic problems. We extend the analysis to the case of convexconcave stochastic saddle point problems, and present (in our opinion highly encouraging) results of numerical experiments.
Tail Bounds for Occupancy and the Satisfiability Threshold Conjecture
, 1995
"... The classical occupancy problem is concerned with studying the number of empty bins resulting from a random allocation of m balls to n bins. We provide a series of tail bounds on the distribution of the number of empty bins. These tail bounds should find application in randomized algorithms and prob ..."
Abstract

Cited by 97 (1 self)
 Add to MetaCart
The classical occupancy problem is concerned with studying the number of empty bins resulting from a random allocation of m balls to n bins. We provide a series of tail bounds on the distribution of the number of empty bins. These tail bounds should find application in randomized algorithms and probabilistic analysis. Our motivating application is the following wellknown conjecture on threshold phenomenon for the satisfiability problem. Consider random 3SAT formulas with cn clauses over n variables, where each clause is chosen uniformly and independently from the space of all clauses of size 3. It has been conjectured that there is a sharp threshold for satisfiability at c ß 4:2. We provide a strong upper bound on the value of c , showing that for c ? 4:758 a random 3SAT formula is unsatisfiable with high probability. This result is based on a structural property, possibly of independent interest, whose proof needs several applications of the occupancy tail bounds. Supporte...