Results 1  10
of
49
Selfish Routing and the Price of Anarchy
, 2005
"... Abstract Selfish routing is a classical mathematical model of how selfinterested users might route traffic through a congested network. The outcome of selfish routing is generally inefficient, in that it fails to optimize natural objective functions. The price of anarchy is a quantitative measure o ..."
Abstract

Cited by 175 (12 self)
 Add to MetaCart
Abstract Selfish routing is a classical mathematical model of how selfinterested users might route traffic through a congested network. The outcome of selfish routing is generally inefficient, in that it fails to optimize natural objective functions. The price of anarchy is a quantitative measure of this inefficiency. We survey recent work that analyzes the price of anarchy of selfish routing. We also describe related results on bounding the worstpossible severity of a phenomenon called Braess's Paradox, and on three techniques for reducing the price of anarchy of selfish routing. This survey concentrates on the contributions of the author's PhD thesis, but also discusses several more recent results in the area.
The price of routing unsplittable flow
 In Proc. 37th Symp. Theory of Computing (STOC
, 2005
"... The essence of the routing problem in real networks is that the traffic demand from a source to destination must be satisfied by choosing a single path between source and destination. The splittable version of this problem is when demand can be satisfied by many paths, namely a flow from source to d ..."
Abstract

Cited by 106 (4 self)
 Add to MetaCart
The essence of the routing problem in real networks is that the traffic demand from a source to destination must be satisfied by choosing a single path between source and destination. The splittable version of this problem is when demand can be satisfied by many paths, namely a flow from source to destination. The unsplittable, or discrete version of the problem is more realistic yet is more complex from the algorithmic point of view; in some settings optimizing such unsplittable traffic flow is computationally intractable. In this paper, we assume this more realistic unsplittable model, and investigate the ”price of anarchy”, or deterioration of network performance measured in total traffic latency under the selfish user behavior. We show that for linear edge latency functions the price of anarchy is exactly 2.618 for weighted demand and exactly 2.5 for unweighted demand. These results are easily extended to (weighted or unweighted) atomic ”congestion games”, where paths are replaced by general subsets. We also show that for polynomials of degree d edge latency functions the price of anarchy is dΘ(d). Our results hold also for mixed strategies. Previous results of Roughgarden and Tardos showed that for linear edge latency functions the price of anarchy is exactly 4 3 under the assumption that each user controls only a negligible fraction of the overall traffic (this result also holds for the splittable case). Note that under the assumption of negligible traffic pure and mixed strategies are equivalent and also splittable and unsplittable models are equivalent. 1
Pricing network edges for heterogeneous selfish users
 Proc. of STOC
, 2003
"... We study the negative consequences of selfish behavior in a congested network and economic means of influencing such behavior. We consider the model of selfish routing defined by Wardrop [30] and studied in a computer science context by Roughgarden and Tardos [26]. In this model, the latency experie ..."
Abstract

Cited by 88 (10 self)
 Add to MetaCart
We study the negative consequences of selfish behavior in a congested network and economic means of influencing such behavior. We consider the model of selfish routing defined by Wardrop [30] and studied in a computer science context by Roughgarden and Tardos [26]. In this model, the latency experienced by network traffic on an edge of the network is a function of the edge congestion, and network users are assumed to selfishly route traffic on minimumlatency paths. The quality of a routing of traffic is measured by the sum of travel times (the total latency). It is well known that the outcome of selfish routing (a Nash equilibrium) does not minimize the total latency and can be improved upon with coordination. An ancient strategy for improving the selfish solution is the principle of marginal cost pricing, which asserts that on each edge of the network, each network user on the edge should pay a tax offsetting the congestion effects caused by its presence. By pricing network edges according to this principle, the inefficiency of selfish routing can always be eradicated. This result, while fundamental, assumes a very strong homogeneity property: all network users are assumed to trade off time and money in an identical way. The guarantee also ignores both the algorithmic
Coordination mechanisms
 PROCEEDINGS OF THE 31ST INTERNATIONAL COLLOQUIUM ON AUTOMATA, LANGUAGES AND PROGRAMMING, IN: LECTURE NOTES IN COMPUTER SCIENCE
, 2004
"... We introduce the notion of coordination mechanisms to improve the performance in systems with independent selfish and noncolluding agents. The quality of a coordination mechanism is measured by its price of anarchy—the worstcase performance of a Nash equilibrium over the (centrally controlled) soc ..."
Abstract

Cited by 42 (5 self)
 Add to MetaCart
We introduce the notion of coordination mechanisms to improve the performance in systems with independent selfish and noncolluding agents. The quality of a coordination mechanism is measured by its price of anarchy—the worstcase performance of a Nash equilibrium over the (centrally controlled) social optimum. We give upper and lower bounds for the price of anarchy for selfish task allocation and congestion games.
Optimal mechanism design and money burning
 STOC ’08
, 2008
"... Mechanism design is now a standard tool in computer science for aligning the incentives of selfinterested agents with the objectives of a system designer. There is, however, a fundamental disconnect between the traditional application domains of mechanism design (such as auctions) and those arising ..."
Abstract

Cited by 37 (12 self)
 Add to MetaCart
Mechanism design is now a standard tool in computer science for aligning the incentives of selfinterested agents with the objectives of a system designer. There is, however, a fundamental disconnect between the traditional application domains of mechanism design (such as auctions) and those arising in computer science (such as networks): while monetary transfers (i.e., payments) are essential for most of the known positive results in mechanism design, they are undesirable or even technologically infeasible in many computer systems. Classical impossibility results imply that the reach of mechanisms without transfers is severely limited. Computer systems typically do have the ability to reduce service quality—routing systems can drop or delay traffic, scheduling protocols can delay the release of jobs, and computational payment schemes can require computational payments from users (e.g., in spamfighting systems). Service degradation is tantamount to requiring that users burn money, and such “payments ” can be used to influence the preferences of the agents at a cost of degrading the social surplus. We develop a framework for the design and analysis of moneyburning mechanisms to maximize the residual surplus— the total value of the chosen outcome minus the payments required. Our primary contributions are the following. • We define a general template for priorfree optimal mechanism design that explicitly connects Bayesian optimal mechanism design, the dominant paradigm in economics, with worstcase analysis. In particular, we establish a general and principled way to identify appropriate performance benchmarks in priorfree mechanism design. • For general singleparameter agent settings, we char
(Almost) optimal coordination mechanisms for unrelated maching scheduling
 IN 18TH ACMSIAM SYMP. ON DISCRETE ALGORITHMS (SODA
, 2008
"... We investigate the influence of different algorithmic choices on the approximation ratio in selfish scheduling. Our goal is to design local policies that minimize the inefficiency of resulting equilibria. In particular, we design optimal coordination mechanisms for unrelated machine scheduling, and ..."
Abstract

Cited by 21 (4 self)
 Add to MetaCart
We investigate the influence of different algorithmic choices on the approximation ratio in selfish scheduling. Our goal is to design local policies that minimize the inefficiency of resulting equilibria. In particular, we design optimal coordination mechanisms for unrelated machine scheduling, and improve the known approximation ratio from Θ(m) to Θ(log m), where m is the number of machines. A local policy for each machine orders the set of jobs assigned to it only based on parameters of those jobs. A strongly local policy only uses the processing time of jobs on the the same machine. We prove that the approximation ratio of any set of strongly local ordering policies in equilibria is at least Ω(m). In particular, it implies that the approximation ratio of a greedy shortestfirst algorithm for machine scheduling is at least Ω(m). This closes the gap between the known lower and upper bounds for this problem, and answers an open question raised by Ibarra and Kim [16], and Davis and Jaffe [10]. We then design a local ordering policy with the approximation ratio of Θ(log m) in equilibria, and prove that this policy is optimal among all local ordering policies. This policy orders the jobs in the nondecreasing order of their inefficiency, i.e, the ratio between the processing time on that machine over the minimum processing time. Finally, we show that best responses of players for the inefficiencybased policy may not converge to a pure Nash equilibrium, and present a Θ(log² m) policy for which we can prove fast convergence of best responses to pure Nash equilibria.
Stackelberg thresholds in network routing games or the value of altruism
 EC'07
, 2007
"... Noncooperative network routing games are a natural model of users trying to selfishly route flow through a network in order to minimize their own delays. It is well known that the solution resulting from this selfish routing (called the Nash equilibrium) can have social cost strictly higher than the ..."
Abstract

Cited by 16 (0 self)
 Add to MetaCart
Noncooperative network routing games are a natural model of users trying to selfishly route flow through a network in order to minimize their own delays. It is well known that the solution resulting from this selfish routing (called the Nash equilibrium) can have social cost strictly higher than the cost of the optimum solution. One way to improve the quality of the resulting solution is to centrally control a fraction of the flow. A natural problem for the network administrator then is to route the centrally controlled flow in such a way that the overall cost of the solution is minimized after the remaining fraction has routed itself selfishly. This problem falls in the class of wellstudied Stackelberg routing games. We consider the scenario where the network administrator wants the final solution to be (strictly) better than the Nash equilibrium. In other words, she wants to control enough flow such that the cost of the resulting solution is strictly less than the cost of the Nash equilibrium. We call the minimum fraction of users that must be centrally routed to improve the quality of the resulting solution the Stackelberg threshold. We give a closed form expression for the Stackelberg threshold for parallel links networks with linear latency functions. The expression is in terms of Nash equilibrium flows and optimum flows. It turns out that the Stackelberg threshold is the minimum of Nash flows on links which have more optimum flow than Nash flow. Using our approach to characterize the Stackelberg thresholds, we are able to give a simpler proof of an earlier result which finds the minimum fraction required to be centrally controlled to induce an optimum solution.