Results 1  10
of
117
On the Combinatorial and Algebraic Complexity of Quantifier Elimination
, 1996
"... In this paper, a new algorithm for performing quantifier elimination from first order formulas over real closed fields is given. This algorithm improves the complexity of the asymptotically fastest algorithm for this problem, known to this date. A new feature of this algorithm is that the role of th ..."
Abstract

Cited by 197 (28 self)
 Add to MetaCart
In this paper, a new algorithm for performing quantifier elimination from first order formulas over real closed fields is given. This algorithm improves the complexity of the asymptotically fastest algorithm for this problem, known to this date. A new feature of this algorithm is that the role of the algebraic part (the dependence on the degrees of the input polynomials) and the combinatorial part (the dependence on the number of polynomials) are separated. Another new feature is that the degrees of the polynomials in the equivalent quantifierfree formula that is output, are independent of the number of input polynomials. As special cases of this algorithm, new and improved algorithms for deciding a sentence in the first order theory over real closed fields, and also for solving the existential problem in the first order theory over real closed fields, are obtained.
A Gröbner free alternative for polynomial system solving
 Journal of Complexity
, 2001
"... Given a system of polynomial equations and inequations with coefficients in the field of rational numbers, we show how to compute a geometric resolution of the set of common roots of the system over the field of complex numbers. A geometric resolution consists of a primitive element of the algebraic ..."
Abstract

Cited by 82 (17 self)
 Add to MetaCart
Given a system of polynomial equations and inequations with coefficients in the field of rational numbers, we show how to compute a geometric resolution of the set of common roots of the system over the field of complex numbers. A geometric resolution consists of a primitive element of the algebraic extension defined by the set of roots, its minimal polynomial and the parametrizations of the coordinates. Such a representation of the solutions has a long history which goes back to Leopold Kronecker and has been revisited many times in computer algebra. We introduce a new generation of probabilistic algorithms where all the computations use only univariate or bivariate polynomials. We give a new codification of the set of solutions of a positive dimensional algebraic variety relying on a new global version of Newton’s iterator. Roughly speaking the complexity of our algorithm is polynomial in some kind of degree of the system, in its height, and linear in the complexity of evaluation
LOWER BOUNDS FOR DIOPHANTINE APPROXIMATIONS
, 1996
"... We introduce a subexponential algorithm for geometric solving of multivariate polynomial equation systems whose bit complexity depends mainly on intrinsic geometric invariants of the solution set. ¿From this algorithm, we derive a new procedure for the decision of consistency of polynomial equation ..."
Abstract

Cited by 62 (24 self)
 Add to MetaCart
We introduce a subexponential algorithm for geometric solving of multivariate polynomial equation systems whose bit complexity depends mainly on intrinsic geometric invariants of the solution set. ¿From this algorithm, we derive a new procedure for the decision of consistency of polynomial equation systems whose bit complexity is subexponential, too. As a byproduct, we analyze the division of a polynomial modulo a reduced complete intersection ideal and from this, we obtain an intrinsic lower bound for the logarithmic height of diophantine approximations to a given solution of a zero–dimensional polynomial equation system. This result represents a multivariate version of Liouville’s classical theorem on approximation of algebraic numbers by rationals. A special feature of our procedures is their polynomial character with respect to the mentioned geometric invariants when instead of bit operations only arithmetic operations are counted at unit cost. Technically our paper relies on the use of straight–line programs as a data structure for the encoding of polynomials, on a new symbolic application of Newton’s algorithm to the Implicit Function Theorem and on a special, basis independent trace formula for affine Gorenstein algebras.
When polynomial equation systems can be "solved" fast?
 IN PROC. 11TH INTERNATIONAL SYMPOSIUM APPLIED ALGEBRA, ALGEBRAIC ALGORITHMS AND ERRORCORRECTING CODES, AAECC11
, 1995
"... We present a new method for solving symbolically zerodimensional polynomial equation systems in the affine and toric case. The main feature of our method is the use of an alternative data structure: arithmetic networks and straightline programs with FOR gates. For sequential time complexity measu ..."
Abstract

Cited by 61 (19 self)
 Add to MetaCart
We present a new method for solving symbolically zerodimensional polynomial equation systems in the affine and toric case. The main feature of our method is the use of an alternative data structure: arithmetic networks and straightline programs with FOR gates. For sequential time complexity measured by the size of these networks we obtain the following result: it is possible to solve any affine or toric zerodimensional equation system in nonuniform sequential time which is polynomial in the length of the input description and the &quot;geometric degree &quot; of the equation system. Here, the input is thought to be given by a straightline program (or alternatively in sparse representation), and the length of the input is measured by number of variables, degree of equations and size of the program (or sparsity of the equations). Geometric degree has to be adequately defined. It is always bounded by the algebraiccombinatoric &quot;B'ezout number &quot; of the system which is given by the Hilbert function of a suitable homogeneous ideal. However, in many important cases, the value of the geometric degree is much smaller than
Straightline programs in geometric elimination theory
 J. Pure Appl. Algebra
, 1998
"... Dedicated to Volker Strassen for his work on complexity We present a new method for solving symbolically zero–dimensional polynomial equation systems in the affine and toric case. The main feature of our method is the use of problem adapted data structures: arithmetic networks and straight–line prog ..."
Abstract

Cited by 57 (14 self)
 Add to MetaCart
Dedicated to Volker Strassen for his work on complexity We present a new method for solving symbolically zero–dimensional polynomial equation systems in the affine and toric case. The main feature of our method is the use of problem adapted data structures: arithmetic networks and straight–line programs. For sequential time complexity measured by network size we obtain the following result: it is possible to solve any affine or toric zero–dimensional equation system in non–uniform sequential time which is polynomial in the length of the input description and the “geometric degree ” of the equation system. Here, the input is thought to be given by a straight–line program (or alternatively in sparse representation), and the length of the input is measured by number of variables, degree of equations and size of the program (or sparsity of the equations). The geometric degree of the input system has to be adequately defined. It is always bounded by the algebraic–combinatoric “Bézout number ” of the system which is given by the Hilbert function of a suitable homogeneous ideal. However, in many important cases, the value of the geometric
Computing the equidimensional decomposition of an algebraic closed set by means of lifting fibers
 J. Complexity
, 2000
"... We present a new probabilistic method for solving systems of polynomial equations and inequations. Our algorithm computes the equidimensional decomposition of the Zariski closure of the solution set of such systems. Each equidimensional component is encoded by a generic fiber, that is a finite set o ..."
Abstract

Cited by 56 (2 self)
 Add to MetaCart
We present a new probabilistic method for solving systems of polynomial equations and inequations. Our algorithm computes the equidimensional decomposition of the Zariski closure of the solution set of such systems. Each equidimensional component is encoded by a generic fiber, that is a finite set of points obtained from the intersection of the component with a generic transverse affine subspace. Our algorithm is incremental in the number of equations to be solved. Its complexity is mainly cubic in the maximum of the degrees of the solution sets of the intermediate systems counting multiplicities. Our method is designed for coefficient fields having characteristic zero or big enough with respect to the number of solutions. If the base field is the field of the rational numbers then the resolution is first performed modulo a random prime number after we have applied a random change of coordinates. Then we search for coordinates with small integers and lift the solutions up to the rational numbers. Our implementation is available within our package Kronecker from version 0.166, which is written in the Magma computer algebra system. 1
On the complexity of numerical analysis
 IN PROC. 21ST ANN. IEEE CONF. ON COMPUTATIONAL COMPLEXITY (CCC ’06
, 2006
"... We study two quite different approaches to understanding the complexity of fundamental problems in numerical analysis: • The BlumShubSmale model of computation over the reals. • A problem we call the “Generic Task of Numerical Computation, ” which captures an aspect of doing numerical computation ..."
Abstract

Cited by 45 (5 self)
 Add to MetaCart
We study two quite different approaches to understanding the complexity of fundamental problems in numerical analysis: • The BlumShubSmale model of computation over the reals. • A problem we call the “Generic Task of Numerical Computation, ” which captures an aspect of doing numerical computation in floating point, similar to the “long exponent model ” that has been studied in the numerical computing community. We show that both of these approaches hinge on the question of understanding the complexity of the following problem, which we call PosSLP: Given a divisionfree straightline program producing an integer N, decide whether N> 0. • In the BlumShubSmale model, polynomial time computation over the reals (on discrete inputs) is polynomialtime equivalent to PosSLP, when there are only algebraic constants. We conjecture that using transcendental constants provides no additional power, beyond nonuniform reductions to PosSLP, and we present some preliminary results supporting this conjecture. • The Generic Task of Numerical Computation is also polynomialtime equivalent to PosSLP. We prove that PosSLP lies in the counting hierarchy. Combining this with work of Tiwari, we obtain that the Euclidean Traveling Salesman Problem lies in the counting hierarchy – the previous best upper bound for this important problem (in terms of classical complexity classes) being PSPACE. In the course of developing the context for our results on arithmetic circuits, we present some new observations on the complexity of ACIT: the Arithmetic Circuit Identity Testing problem. In particular, we show that if n! is not ultimately easy, then ACIT has subexponential complexity.
Geometric Complexity Theory I: An Approach to the P. vs. NP and related problems
, 2001
"... We suggest an approach based on geometric invariant theory to the fundamentallower bound problems in complexity theory concerning formula and circuit size. Specifically, ..."
Abstract

Cited by 44 (12 self)
 Add to MetaCart
We suggest an approach based on geometric invariant theory to the fundamentallower bound problems in complexity theory concerning formula and circuit size. Specifically,
Superpolynomial lower bounds for monotone span programs
, 1996
"... In this paper we obtain the first superpolynomial lower bounds for monotone span programs computing explicit functions. The best previous lower bound was Ω(n 5/2) by Beimel, Gál, Paterson [BGP]; our proof exploits a general combinatorial lower bound criterion from that paper. Our lower bounds are ba ..."
Abstract

Cited by 44 (6 self)
 Add to MetaCart
In this paper we obtain the first superpolynomial lower bounds for monotone span programs computing explicit functions. The best previous lower bound was Ω(n 5/2) by Beimel, Gál, Paterson [BGP]; our proof exploits a general combinatorial lower bound criterion from that paper. Our lower bounds are based on an analysis of Paleytype bipartite graphs via Weil’s character sum estimates. We prove an n Ω(log n / log log n) lower bound for the size of monotone span programs for the clique problem. Our results give the first superpolynomial lower bounds for linear secret sharing schemes. We demonstrate the surprising power of monotone span programs by exhibiting a function computable in this model in linear size while requiring superpolynomial size monotone circuits and exponential size monotone formulae. We also show that the perfect matching function can be computed by polynomial size (nonmonotone) span programs over arbitrary fields.
Structure and Importance of LogspaceMODClasses
, 1992
"... . We refine the techniques of Beigel, Gill, Hertrampf [4] who investigated polynomial time counting classes, in order to make them applicable to the case of logarithmic space. We define the complexity classes MOD k L and demonstrate their significance by proving that all standard problems of linear ..."
Abstract

Cited by 41 (1 self)
 Add to MetaCart
. We refine the techniques of Beigel, Gill, Hertrampf [4] who investigated polynomial time counting classes, in order to make them applicable to the case of logarithmic space. We define the complexity classes MOD k L and demonstrate their significance by proving that all standard problems of linear algebra over the finite rings Z/kZ are complete for these classes. We then define new complexity classes LogFew and LogFewNL and identify them as adequate logspace versions of Few and FewP. We show that LogFewNL is contained in MODZ k L and that LogFew is contained in MOD k L for all k. Also an upper bound for L #L in terms of computation of integer determinants is given from which we conclude that all logspace counting classes are contained in NC 2 . 1 Introduction Valiant [21] defined the class #P of functions f such that there is a nondeterministic polynomial time Turing machine which, on input x, has exactly f(x) accepting computation paths. Many complexity classes in the area betw...