Results 1  10
of
34
A New Rounding Procedure for the Assignment Problem with Applications to Dense Graph Arrangement Problems
, 2001
"... We present a randomized procedure for rounding fractional perfect matchings to (integral) matchings. If the original fractional matching satis es any linear inequality, then with high probability, the new matching satis es that linear inequality in an approximate sense. This extends the wellkn ..."
Abstract

Cited by 77 (3 self)
 Add to MetaCart
We present a randomized procedure for rounding fractional perfect matchings to (integral) matchings. If the original fractional matching satis es any linear inequality, then with high probability, the new matching satis es that linear inequality in an approximate sense. This extends the wellknown LP rounding procedure of Raghavan and Thompson, which is usually used to round fractional solutions of linear programs.
The probabilistic method yields deterministic parallel algorithms
 Journal of Computer and System Sciences
, 1989
"... We present a technique for converting RNC algorithms into NC algorithms. Our approach is based on a parallel implementation of the method of conditional probabilities. This method was used to convert probabilistic proofs of existence of combinatorial structures into polynomial time deterministic alg ..."
Abstract

Cited by 53 (7 self)
 Add to MetaCart
We present a technique for converting RNC algorithms into NC algorithms. Our approach is based on a parallel implementation of the method of conditional probabilities. This method was used to convert probabilistic proofs of existence of combinatorial structures into polynomial time deterministic algorithms. It has the apparent drawback of being extremely sequential in nature. We show certain general conditions under which it is possible to use this technique for devising deterministic parallel algorithms. We use our technique to devise an NC algorithm for the set balancing problem. This problem turns out to be a useful tool for parallel algorithms. Using our derandomization method and the set balancing algorithm, we provide an NC algorithm for the lattice approximation problem. We also use the lattice approximation problem to bootstrap the set balancing algorithm, and the result is a more processor efficient algorithm. The set balancing algorithm also yields an NC algorithm for nearoptimal edge coloring of simple graphs. Our methods also extend to the parallelization of various algorithms in computational geometry that rely upon the random sampling technique of Clarkson. Finally, our methods apply to constructing certain combinatorial structures, e.g. ...
An NC Algorithm for Minimum Cuts
 IN PROCEEDINGS OF THE 25TH ANNUAL ACM SYMPOSIUM ON THEORY OF COMPUTING
"... We show that the minimum cut problem for weighted undirected graphs can be solved in NC using three separate and independently interesting results. The first is an (m 2 =n)processor NC algorithm for finding a (2 + ffl)approximation to the minimum cut. The second is a randomized reduction from ..."
Abstract

Cited by 46 (3 self)
 Add to MetaCart
We show that the minimum cut problem for weighted undirected graphs can be solved in NC using three separate and independently interesting results. The first is an (m 2 =n)processor NC algorithm for finding a (2 + ffl)approximation to the minimum cut. The second is a randomized reduction from the minimum cut problem to the problem of obtaining a (2 + ffl)approximation to the minimum cut. This reduction involves a natural combinatorial SetIsolation Problem that can be solved easily in RNC. The third result is a derandomization of this RNC solution that requires a combination of two widely used tools: pairwise independence and random walks on expanders. We believe that the setisolation approach will prove useful in other derandomization problems. The techniques extend to two related problems: we describe NC algorithms finding minimum kway cuts for any constant k and finding all cuts of value within any constant factor of the minimum. Another application of these techni...
On the Hardness of Graph Isomorphism
 SIAM J. COMPUT
"... We show that the graph isomorphism problem is hard under DLOGTIME uniform AC0 manyone reductions for the complexity classes NL, PL (probabilistic logarithmic space) for every logarithmic space modular class ModkL and for the class DET of problems NC¹ reducible to the determinant. These are the stro ..."
Abstract

Cited by 32 (1 self)
 Add to MetaCart
We show that the graph isomorphism problem is hard under DLOGTIME uniform AC0 manyone reductions for the complexity classes NL, PL (probabilistic logarithmic space) for every logarithmic space modular class ModkL and for the class DET of problems NC¹ reducible to the determinant. These are the strongest known hardness results for the graph isomorphism problem and imply a randomized logarithmic space reduction from the perfect matching problem to graph isomorphism. We also investigate hardness results for the graph automorphism problem.
Reducing Randomness Via Irrational Numbers
 In Proceedings of the TwentyNinth Annual ACM Symposium on Theory of Computing
, 1997
"... . We propose a general methodology for testing whether a given polynomial with integer coefficients is identically zero. The methodology evaluates the polynomial at efficiently computable approximations of suitable irrational points. In contrast to the classical technique of DeMillo, Lipton, Schwart ..."
Abstract

Cited by 23 (0 self)
 Add to MetaCart
. We propose a general methodology for testing whether a given polynomial with integer coefficients is identically zero. The methodology evaluates the polynomial at efficiently computable approximations of suitable irrational points. In contrast to the classical technique of DeMillo, Lipton, Schwartz, and Zippel, this methodology can decrease the error probability by increasing the precision of the approximations instead of using more random bits. Consequently, randomized algorithms that use the classical technique can generally be improved using the new methodology. To demonstrate the methodology, we discuss two nontrivial applications. The first is to decide whether a graph has a perfect matching in parallel. Our new NC algorithm uses fewer random bits while doing less work than the previously best NC algorithm by Chari, Rohatgi, and Srinivasan. The second application is to test the equality of two multisets of integers. Our new algorithm improves upon the previously best algorithms ...
Isolation, Matching, and Counting: Uniform and Nonuniform Upper Bounds
 Journal of Computer and System Sciences
, 1998
"... We show that the perfect matching problem is in the complexity class SPL (in the nonuniform setting). This provides a better upper bound on the complexity of the matching problem, as well as providing motivation for studying the complexity class SPL. Using similar techniques, we show that counting t ..."
Abstract

Cited by 22 (4 self)
 Add to MetaCart
We show that the perfect matching problem is in the complexity class SPL (in the nonuniform setting). This provides a better upper bound on the complexity of the matching problem, as well as providing motivation for studying the complexity class SPL. Using similar techniques, we show that counting the number of accepting paths of a nondeterministic logspace machine can be done in NL/poly, if the number of paths is small. This clarifies the complexity of the class LogFew (defined and studied in [BDHM91]). Using derandomization techniques, we then improve this to show that this counting problem is in NL. Determining if our other theorems hold in the uniform setting remains an The material in this paper appeared in preliminary form in papers in the Proceedings of the IEEE Conference on Computational Complexity, 1998, and in the Proceedings of the Workshop on Randomized Algorithms, Brno, 1998. y Supported in part by NSF grants CCR9509603 and CCR9734918. z Supported in part by the ...
Complexity Issues in Discrete Hopfield Networks
, 1994
"... We survey some aspects of the computational complexity theory of discretetime and discretestate Hopfield networks. The emphasis is on topics that are not adequately covered by the existing survey literature, most significantly: 1. the known upper and lower bounds for the convergence times of Hopfi ..."
Abstract

Cited by 18 (4 self)
 Add to MetaCart
We survey some aspects of the computational complexity theory of discretetime and discretestate Hopfield networks. The emphasis is on topics that are not adequately covered by the existing survey literature, most significantly: 1. the known upper and lower bounds for the convergence times of Hopfield nets (here we consider mainly worstcase results); 2. the power of Hopfield nets as general computing devices (as opposed to their applications to associative memory and optimization); 3. the complexity of the synthesis ("learning") and analysis problems related to Hopfield nets as associative memories. Draft chapter for the forthcoming book The Computational and Learning Complexity of Neural Networks: Advanced Topics (ed. Ian Parberry).
(De)randomized Construction of Small Sample Spaces in NC
, 1994
"... Koller and Megiddo introduced the paradigm of constructing compact distributions that satisfy a given set of constraints, and showed how it can be used to efficiently derandomize certain types of algorithm. In this paper, we significantly extend their resdts in two ways. First, we show how their app ..."
Abstract

Cited by 17 (0 self)
 Add to MetaCart
Koller and Megiddo introduced the paradigm of constructing compact distributions that satisfy a given set of constraints, and showed how it can be used to efficiently derandomize certain types of algorithm. In this paper, we significantly extend their resdts in two ways. First, we show how their approach can be applied to deal with more general expectation constraints. More importantly, we provide the first parallel (Ne) algorithm for constructing a compact distribution that satisfies the constraints up to a small relative error. This algorithm deals with constraints over any event that can be verified by finite automata, including all independence constraints as well as constraints over events relating to the parity OT sum of a certain set of variables. OUT construction relies on a new and independently interesting parallel algorithm for converting a solution to a linear system into an almost basic approximate solution to the same system. We use these techniques in the first AfC derandomization of an algorithm for constructing large independent sets in duniform hypergraphs for arbitrary d. We also show how the linear programming perspective suggests new proof techniques which might be useful in general probabilistic analysis.
Distributed approximate matching
, 2007
"... We consider distributed algorithms for approximate maximum matching on general graphs. Our main result is a randomized (4 + ɛ)approximation distributed algorithm for weighted maximum matching, whose running time is O(log n) for any constant ɛ> 0, where n is the number of nodes in the graph. In addi ..."
Abstract

Cited by 16 (2 self)
 Add to MetaCart
We consider distributed algorithms for approximate maximum matching on general graphs. Our main result is a randomized (4 + ɛ)approximation distributed algorithm for weighted maximum matching, whose running time is O(log n) for any constant ɛ> 0, where n is the number of nodes in the graph. In addition, we consider the dynamic case, where nodes are inserted and deleted one at a time. For unweighted dynamic graphs, we give an algorithm that maintains a (1 + ɛ)approximation in O(1/ɛ) time for each node insertion or deletion. For weighted dynamic graphs we give a constantfactor approximation algorithm that runs in constant time for each insertion or deletion.