Results 1  10
of
65
Coalition Structure Generation with Worst Case Guarantees
, 1999
"... Coalition formation is a key topic in multiagent systems. One may prefer a coalition structure that maximizes the sum of the values of the coalitions, but often the number of coalition structures is too large to allow exhaustive search for the optimal one. Furthermore, finding the optimal coalition ..."
Abstract

Cited by 208 (10 self)
 Add to MetaCart
Coalition formation is a key topic in multiagent systems. One may prefer a coalition structure that maximizes the sum of the values of the coalitions, but often the number of coalition structures is too large to allow exhaustive search for the optimal one. Furthermore, finding the optimal coalition structure is NPcomplete. But then, can the coalition structure found via a partial search be guaranteed to be within a bound from optimum? We show that none of the previous coalition structure generation algorithms can establish any bound because they search fewer nodes than a threshold that we show necessary for establishing a bound. We present an algorithm that establishes a tight bound within this minimal amount of search, and show that any other algorithm would have to search strictly more. The fraction of nodes needed to be searched approaches zero as the number of agents grows. If additional time remains, our anytime algorithm searches further, and establishes a progressively lower tight bound. Surprisingly, just searching one more node drops the bound in half. As desired, our algorithm lowers the bound rapidly early on, and exhibits diminishing returns to computation. It also significantly outperforms its obvious contenders. Finally, we show how to distribute the desired
Simple heuristics for unit disk graphs
 NETWORKS
, 1995
"... Unit disk graphs are intersection graphs of circles of unit radius in the plane. We present simple and provably good heuristics for a number of classical NPhard optimization problems on unit disk graphs. The problems considered include maximum independent set, minimum vertex cover, minimum coloring ..."
Abstract

Cited by 125 (6 self)
 Add to MetaCart
Unit disk graphs are intersection graphs of circles of unit radius in the plane. We present simple and provably good heuristics for a number of classical NPhard optimization problems on unit disk graphs. The problems considered include maximum independent set, minimum vertex cover, minimum coloring and minimum dominating set. We also present an online coloring heuristic which achieves a competitive ratio of 6 for unit disk graphs. Our heuristics do not need a geometric representation of unit disk graphs. Geometric representations are used only in establishing the performance guarantees of the heuristics. Several of our approximation algorithms can be extended to intersection graphs of circles of arbitrary radii in the plane, intersection graphs of regular polygons, and to intersection graphs of higher dimensional regular objects.
On syntactic versus computational views of approximability
 SIAM JOURNAL ON COMPUTING
, 1999
"... We attempt to reconcile the two distinct views of approximation classes: syntactic and computational. Syntactic classes such as MAX SNP permit structural results and have natural complete problems, while computational classes such as APX allow us to work with classes of problems whose approximabilit ..."
Abstract

Cited by 117 (12 self)
 Add to MetaCart
We attempt to reconcile the two distinct views of approximation classes: syntactic and computational. Syntactic classes such as MAX SNP permit structural results and have natural complete problems, while computational classes such as APX allow us to work with classes of problems whose approximability is wellunderstood. Our results provide a syntactic characterization of computational classes, and give a computational framework for syntactic classes. We compare the syntactically defined class MAX SNP with the computationally defined class APX, and show that every problem in APX can be “placed ” (i.e. has approximation preserving reduction to a problem) in MAX SNP. Our methods introduce a general technique for creating approximationpreserving reductions which show that any “well ” approximable problem can be reduced in an approximationpreserving manner to a problem which is hard to approximate to corresponding factors. We demonstrate this technique by applying it to the classes RMAX(2) and MIN F+Π2(1)which have the clique problem and the set cover problem, respectively, as complete problems. We use the syntactic nature of MAX SNP to define a general paradigm, nonoblivious local search, useful for developing simple yet efficient approximation algorithms. We show that such algorithms can find good approximations for all MAX SNP problems, yielding approximation ratios comparable to the bestknown for a variety of specific MAX SNPhard problems. Nonoblivious local search provably outperforms standard local search in both the degree of approximation achieved and the efficiency of the resulting algorithms.
eMediator: A Next Generation Electronic Commerce Server
 Computational Intelligence
, 2002
"... This paper presents eMediator, an electronic commerce server prototype that demonstrates ways in which algorithmic support and gametheoretic incentive engineering can jointly improve the efficiency of ecommerce. eAuctionHouse, the configurable auction server, includes a variety of generalized combi ..."
Abstract

Cited by 107 (31 self)
 Add to MetaCart
This paper presents eMediator, an electronic commerce server prototype that demonstrates ways in which algorithmic support and gametheoretic incentive engineering can jointly improve the efficiency of ecommerce. eAuctionHouse, the configurable auction server, includes a variety of generalized combinatorial auctions and exchanges, pricing schemes, bidding languages, mobile agents, and user support for choosing an auction type. We introduce two new logical bidding languages for combinatorial markets: the XOR bidding language and the ORofXORs bidding language. Unlike the traditional OR bidding language, these are fully expressive. They therefore enable the use of the ClarkeGroves pricing mechanism for motivating the bidders to bid truthfully. eAuctionHouse also supports supply/demand curve bidding. eCommitter, the leveled commitment contract optimizer, determines the optimal contract price and decommitting penalties for a variety of leveled commitment contracting mechanisms, taking into account that rational agents will decommit strategically in Nash equilibrium. It also determines the optimal decommitting strategies for any given leveled commitment contract. eExchangeHouse, the safe exchange planner, enables unenforced anonymous exchanges by dividing the exchange into chunks and sequencing those chunks to be delivered safely in alternation between the buyer and the seller.
Improved Approximation Algorithms for the Vertex Cover Problem in Graphs and Hypergraphs
, 1999
"... We obtain improved algorithms for finding small vertex covers in bounded degree graphs and hypergraphs. We use semidefinite programming to relax the problems, and introduce new rounding techniques for these relaxations. On graphs with maximum degree at most Δ, the algorithm achieves a performa ..."
Abstract

Cited by 92 (6 self)
 Add to MetaCart
We obtain improved algorithms for finding small vertex covers in bounded degree graphs and hypergraphs. We use semidefinite programming to relax the problems, and introduce new rounding techniques for these relaxations. On graphs with maximum degree at most Δ, the algorithm achieves a performance ratio of 2  (1  o(1)) 2 ln ln \Delta ln \Delta for large \Delta, which improves the previously known ratio of 2 \Gamma log \Delta+O(1) \Delta obtained by Halldórsson and Radhakrishnan. Using similar techniques, we also present improved approximations for the vertex cover problem in hypergraphs. For kuniform hypergraphs with n vertices, we achieve a ratio of k \Gamma (1 \Gamma o(1)) k ln ln n ln n for large n, and for kuniform hypergraphs with maximum degree at most \Delta, the algorithm achieves a ratio of k \Gamma (1 \Gamma o(1)) k(k\Gamma1) ln ln \Delta ln \Delta for large \Delta. These results considerably improve the previous best ratio of k(1\Gammac=\Delta 1 k\Gamma1 ) for bounded degree kuniform hypergraphs, and k(1 \Gamma c=n k\Gamma1 k ) for general kuniform hypergraphs, both obtained by Krivelevich. Using similar techniques, we also obtain an approximation algorithm for the weighted independent set problem, matching a recent result of Halldórsson.
Finding frequent patterns in a large sparse graph
 SIAM Data Mining Conference
, 2004
"... This paper presents two algorithms based on the horizontal and vertical pattern discovery paradigms that find the connected subgraphs that have a sufficient number of edgedisjoint embeddings in a single large undirected labeled sparse graph. These algorithms use three different methods to determine ..."
Abstract

Cited by 79 (4 self)
 Add to MetaCart
This paper presents two algorithms based on the horizontal and vertical pattern discovery paradigms that find the connected subgraphs that have a sufficient number of edgedisjoint embeddings in a single large undirected labeled sparse graph. These algorithms use three different methods to determine the number of the edgedisjoint embeddings of a subgraph that are based on approximate and exact maximum independent set computations and use it to prune infrequent subgraphs. Experimental evaluation on real datasets from various domains show that both algorithms achieve good performance, scale well to sparse input graphs with more than 100,000 vertices, and significantly outperform a previously developed algorithm.
On problems without polynomial kernels
 Lect. Notes Comput. Sci
, 2007
"... Abstract. Kernelization is a strong and widelyapplied technique in parameterized complexity. In a nutshell, a kernelization algorithm, or simply a kernel, is a polynomialtime transformation that transforms any given parameterized instance to an equivalent instance of the same problem, with size an ..."
Abstract

Cited by 65 (9 self)
 Add to MetaCart
Abstract. Kernelization is a strong and widelyapplied technique in parameterized complexity. In a nutshell, a kernelization algorithm, or simply a kernel, is a polynomialtime transformation that transforms any given parameterized instance to an equivalent instance of the same problem, with size and parameter bounded by a function of the parameter in the input. A kernel is polynomial if the size and parameter of the output are polynomiallybounded by the parameter of the input. In this paper we develop a framework which allows showing that a wide range of FPT problems do not have polynomial kernels. Our evidence relies on hypothesis made in the classical world (i.e. nonparametric complexity), and evolves around a new type of algorithm for classical decision problems, called a distillation algorithm, which might be of independent interest. Using the notion of distillation algorithms, we develop a generic lowerbound engine which allows us to show that a variety of FPT problems, fulfilling certain criteria, cannot have polynomial kernels unless the polynomial hierarchy collapses. These problems include kPath, kCycle, kExact Cycle, kShort Cheap Tour, kGraph Minor Order Test, kCutwidth, kSearch Number, kPathwidth, kTreewidth, kBranchwidth, and several optimization problems parameterized by treewidth or cliquewidth. 1
Greed is Good: Approximating Independent Sets in Sparse and . . .
, 1994
"... ... for short, is one of the ~implest, most efficient, and most thoroughly studied methods for finding independent sets in graphs. We show that it surprisingly achieves a performance ratio of (A+ 2)/3 for approximating independent sets in graphs with degree bounded by A. The analysis directs us tow ..."
Abstract

Cited by 55 (7 self)
 Add to MetaCart
... for short, is one of the ~implest, most efficient, and most thoroughly studied methods for finding independent sets in graphs. We show that it surprisingly achieves a performance ratio of (A+ 2)/3 for approximating independent sets in graphs with degree bounded by A. The analysis directs us towards a simple parallel and distributed algorithm with identical performance, which on constantdegree graphs runs in O(log ” n) time using linear number of processors. We also analyze the Greedy algorithm when run in combination with a fractional relaxation technique of Nemhauser and Trotter, and obtain an improved (2Z + 3)/5 performance ratio on graphs with average degree ~. Finally, we introduce a generally applicable technique for improving the approximation ratios of independent set algorithms, and illustrate it by improving the performance ratio of Greedy for large ∆.
Approximations of Weighted Independent Set and Hereditary Subset Problems
 JOURNAL OF GRAPH ALGORITHMS AND APPLICATIONS
, 2000
"... The focus of this study is to clarify the approximability of weighted versions of the maximum independent set problem. In particular, we report improved performance ratios in boundeddegree graphs, inductive graphs, and general graphs, as well as for the unweighted problem in sparse graphs. Wher ..."
Abstract

Cited by 53 (6 self)
 Add to MetaCart
The focus of this study is to clarify the approximability of weighted versions of the maximum independent set problem. In particular, we report improved performance ratios in boundeddegree graphs, inductive graphs, and general graphs, as well as for the unweighted problem in sparse graphs. Where possible, the techniques are applied to related hereditary subgraph and subset problem, obtaining ratios better than previously reported for e.g. Weighted Set Packing, Longest Common Subsequence, and Independent Set in hypergraphs.
Efficient InterferenceAware TDMA Link Scheduling for Static Wireless Networks
 In ACM MobiCom
, 2006
"... We study efficient link scheduling for a multihop wireless network to maximize its throughput. Efficient link scheduling can greatly reduce the interference effect of closeby transmissions. Unlike the previous studies that often assume a unit disk graph model, we assume that different terminals cou ..."
Abstract

Cited by 47 (7 self)
 Add to MetaCart
We study efficient link scheduling for a multihop wireless network to maximize its throughput. Efficient link scheduling can greatly reduce the interference effect of closeby transmissions. Unlike the previous studies that often assume a unit disk graph model, we assume that different terminals could have different transmission ranges and different interference ranges. In our model, it is also possible that a communication link may not exist due to barriers or is not used by a predetermined routing protocol, while the transmission of a node always result interference to all nonintended receivers within its interference range. Using a mathematical formulation, we develop synchronized TDMA link schedulings that optimize the networking throughput. Specifically, by assuming known link capacities and link traffic loads, we study link scheduling under the RTS/CTS interference model and the protocol interference model with fixed transmission power. For both models, we present both efficient centralized and distributed algorithms that use time slots within a constant factor of the optimum. We also present efficient distributed algorithms whose performances are still comparable with optimum, but with much less communications. Our theoretical results are corroborated by extensive simulation studies.