Results 1 
8 of
8
Engineering an External Memory Minimum Spanning Tree Algorithm
 IN PROC. 3RD IFIP INTL. CONF. ON THEORETICAL COMPUTER SCIENCE
, 2004
"... We develop an external memory algorithm for computing minimum spanning trees. The algorithm is considerably simpler than previously known external memory algorithms for this problem and needs a factor of at least four less I/Os for realistic inputs. Our implementation indicates that this algorithm ..."
Abstract

Cited by 16 (4 self)
 Add to MetaCart
We develop an external memory algorithm for computing minimum spanning trees. The algorithm is considerably simpler than previously known external memory algorithms for this problem and needs a factor of at least four less I/Os for realistic inputs. Our implementation indicates that this algorithm processes graphs only limited by the disk capacity of most current machines in time no more than a factor 2–5 of a good internal algorithm with sufficient memory space.
Experimental Evaluation of a New Shortest Path Algorithm (Extended Abstract)
, 2002
"... We evaluate the practical eciency of a new shortest path algorithm for undirected graphs which was developed by the rst two authors. This algorithm works on the fundamental comparisonaddition model. Theoretically, this new algorithm outperforms Dijkstra's algorithm on sparse graphs for t ..."
Abstract

Cited by 12 (4 self)
 Add to MetaCart
We evaluate the practical eciency of a new shortest path algorithm for undirected graphs which was developed by the rst two authors. This algorithm works on the fundamental comparisonaddition model. Theoretically, this new algorithm outperforms Dijkstra's algorithm on sparse graphs for the allpairs shortest path problem, and more generally, for the problem of computing singlesource shortest paths from !(1) different sources. Our extensive experimental analysis demonstrates that this is also the case in practice. We present results which show the new algorithm to run faster than Dijkstra's on a variety of sparse graphs when the number of vertices ranges from a few thousand to a few million, and when computing singlesource shortest paths from as few as three different sources.
The filterkruskal minimum spanning tree algorithm
, 2009
"... We present FilterKruskal – a simple modification of Kruskal’s algorithm that avoids sorting edges that are “obviously” not in the MST. For arbitrary graphs with random edge weights FilterKruskal runs in time O (m + n lognlog m n, i.e. in linear time for not too sparse graphs. Experiments indicate ..."
Abstract

Cited by 7 (0 self)
 Add to MetaCart
We present FilterKruskal – a simple modification of Kruskal’s algorithm that avoids sorting edges that are “obviously” not in the MST. For arbitrary graphs with random edge weights FilterKruskal runs in time O (m + n lognlog m n, i.e. in linear time for not too sparse graphs. Experiments indicate that the algorithm has very good practical performance over the entire range of edge densities. An equally simple parallelization seems to be the currently best practical algorithm on multicore machines.
A Relational Approach To Optimization Problems
, 1996
"... The main contribution of this thesis is a study of the dynamic programming and greedy strategies for solving combinatorial optimization problems. The study is carried out in the context of a calculus of relations, and generalises previous work by using a loop operator in the imperative programming s ..."
Abstract

Cited by 6 (0 self)
 Add to MetaCart
The main contribution of this thesis is a study of the dynamic programming and greedy strategies for solving combinatorial optimization problems. The study is carried out in the context of a calculus of relations, and generalises previous work by using a loop operator in the imperative programming style for generating feasible solutions, rather than the fold and unfold operators of the functional programming style. The relationship between fold operators and loop operators is explored, and it is shown how to convert from the former to the latter. This fresh approach provides additional insights into the relationship between dynamic programming and greedy algorithms, and helps to unify previously distinct approaches to solving combinatorial optimization problems. Some of the solutions discovered are new and solve problems which had previously proved difficult. The material is illustrated with a selection of problems and solutions that is a mixture of old and new. Another contribution is the invention of a new calculus, called the graph calculus, which is a useful tool for reasoning in the relational calculus and other nonrelational calculi. The graph
Approximation Algorithms for Network Design: A Survey
"... In a typical instance of a network design problem, we are given a directed or undirected graph G = (V,E), nonnegative edgecosts ce for all e ∈ E, and our goal is to find a minimumcost subgraph H of G that satisfies some design criteria. For example, we may wish to find a minimumcost set of edges ..."
Abstract

Cited by 3 (0 self)
 Add to MetaCart
In a typical instance of a network design problem, we are given a directed or undirected graph G = (V,E), nonnegative edgecosts ce for all e ∈ E, and our goal is to find a minimumcost subgraph H of G that satisfies some design criteria. For example, we may wish to find a minimumcost set of edges that induces a connected graph (this is the minimumcost spanning tree problem), or we might want to find a minimumcost set of arcs in a directed graph such that every vertex can reach every other vertex (this is the minimumcost strongly connected subgraph problem). This abstract model for network design problems has a large number of practical applications; the design process of telecommunication and traffic networks, and VLSI chip design are just two examples. Many practically relevant instances of network design problems are NPhard, and thus likely intractable. This survey focuses on approximation algorithms as one possible way of circumventing this impasse. Approximation algorithms are efficient (i.e., they run in polynomialtime), and they compute solutions to a given instance of an optimization problem whose objective values are close to those of the respective optimum solutions. More concretely, most of the problems discussed in this survey are minimization problems. We then say that an algorithm is an αapproximation for a given problem if the ratio of the cost of an approximate solution computed by the algorithm to that of an optimum solution is at most α over all instances. In the
Randomized Minimum Spanning Tree Algorithms Using Exponentially Fewer Random Bits
"... For many fundamental problems there exist randomized algorithms that are asymptotically optimal and are superior to the best known deterministic algorithm. Among these are the minimum spanning tree (MST) problem, the MST sensitivity analysis problem, the parallel connected components and parallel mi ..."
Abstract

Cited by 2 (0 self)
 Add to MetaCart
For many fundamental problems there exist randomized algorithms that are asymptotically optimal and are superior to the best known deterministic algorithm. Among these are the minimum spanning tree (MST) problem, the MST sensitivity analysis problem, the parallel connected components and parallel minimum spanning tree problems, and the local sorting and set maxima problems. (For the first two problems there are provably optimal deterministic algorithms with unknown, and possibly superlinear running times.) One downside of the randomized methods for solving these problems is that they use a number of random bits linear in the size of the input. In this paper we develop some general methods for reducing exponentially the consumption of random bits in comparison based algorithms. In some cases we are able to reduce the number of random bits from linear to nearly constant without affecting the expected running time. Most of our results are obtained by adjusting or reorganizing existing randomized algorithms to work well with a pairwise or O(1)wise independent sampler. The prominent exception — and the main focus of this paper — is a lineartime randomized minimum spanning tree algorithm that is not derived from the well known KargerKleinTarjan algorithm. In many ways it resembles more closely the deterministic minimum spanning tree algorithms based on Soft Heaps. Further,
Fast Energy Optimal Routing in Radio Networks Using Geometric Data Structures \Lambda
, 2002
"... Abstract Given the current position of n nodes in a radio network, we discuss the problem of finding routes between pairs of nodes such that the energy consumption for this communication is minimized. Though this can be done using Dijkstra's algorithm on the complete graph G = (V; V \Theta V) i ..."
Abstract
 Add to MetaCart
Abstract Given the current position of n nodes in a radio network, we discuss the problem of finding routes between pairs of nodes such that the energy consumption for this communication is minimized. Though this can be done using Dijkstra's algorithm on the complete graph G = (V; V \Theta V) in time O \Gamma