Results 1  10
of
636
Solving Polynomial Systems Using a Branch and Prune Approach
 SIAM Journal on Numerical Analysis
, 1997
"... This paper presents Newton, a branch & prune algorithm to find all isolated solutions of a system of polynomial constraints. Newton can be characterized as a global search method which uses intervals for numerical correctness and for pruning the search space early. The pruning in Newton consists ..."
Abstract

Cited by 112 (7 self)
 Add to MetaCart
(Show Context)
This paper presents Newton, a branch & prune algorithm to find all isolated solutions of a system of polynomial constraints. Newton can be characterized as a global search method which uses intervals for numerical correctness and for pruning the search space early. The pruning in Newton consists in enforcing at each node of the search tree a unique local consistency condition, called boxconsistency, which approximates the notion of arcconsistency wellknown in artificial intelligence. Boxconsistency is parametrized by an interval extension of the constraint and can be instantiated to produce the HansenSegupta's narrowing operator (used in interval methods) as well as new operators which are more effective when the computation is far from a solution. Newton has been evaluated on a variety of benchmarks from kinematics, chemistry, combustion, economics, and mechanics. On these benchmarks, it outperforms the interval methods we are aware of and compares well with stateoftheart continuation methods. Limitations of Newton (e.g., a sensitivity to the size of the initial intervals on some problems) are also discussed. Of particular interest is the mathematical and programming simplicity of the method.
Validated Solutions Of Initial Value Problems For Ordinary Differential Equations
, 1996
"... . Compared to standard numerical methods for initial value problems (IVPs) for ordinary differential equations (ODEs), validated methods for IVPs for ODEs have two important advantages: if they return a solution to a problem, then (1) the problem is guaranteed to have a unique solution, and (2) an e ..."
Abstract

Cited by 103 (12 self)
 Add to MetaCart
. Compared to standard numerical methods for initial value problems (IVPs) for ordinary differential equations (ODEs), validated methods for IVPs for ODEs have two important advantages: if they return a solution to a problem, then (1) the problem is guaranteed to have a unique solution, and (2) an enclosure of the true solution is produced. The authors survey Taylor series methods for validated solutions of IVPs for ODEs, describe several such methods in a common framework, and identify areas for future research. Key words. initial value problems, ordinary differential equations, interval arithmetic, Taylor series methods. AMS subject classifications. 65L05, 65G10, 65L60. 1. Introduction. We consider validated numerical methods for the solution of the autonomous initialvalue problems (IVPs) y 0 (t) = f(y); y(t 0 ) = y 0 ; (1.1) where t 2 [t 0 ; T ] for some T ? t 0 . Here t 0 ; T 2 R,f 2 C k\Gamma1 (D), D ` R n is an open set, f : D ! R n , and y 0 2 D. For expositional c...
Efficient solving of large nonlinear arithmetic constraint systems with complex boolean structure
 Journal on Satisfiability, Boolean Modeling and Computation
, 2007
"... In order to facilitate automated reasoning about large Boolean combinations of nonlinear arithmetic constraints involving transcendental functions, we provide a tight integration of recent SAT solving techniques with intervalbased arithmetic constraint solving. Our approach deviates substantially f ..."
Abstract

Cited by 85 (11 self)
 Add to MetaCart
(Show Context)
In order to facilitate automated reasoning about large Boolean combinations of nonlinear arithmetic constraints involving transcendental functions, we provide a tight integration of recent SAT solving techniques with intervalbased arithmetic constraint solving. Our approach deviates substantially from lazy theorem proving approaches in that it directly controls arithmetic constraint propagation from the SAT solver rather than delegating arithmetic decisions to a subordinate solver. Through this tight integration, all the algorithmic enhancements that were instrumental to the enormous performance gains recently achieved in propositional SAT solving carry over smoothly to the rich domain of nonlinear arithmetic constraints. As a consequence, our approach is able to handle large constraint systems with extremely complex Boolean structure, involving Boolean combinations of multiple thousand arithmetic constraints over some thousands of variables.
Safety verification of hybrid systems by constraint propagation based abstraction refinement
, 2005
"... This paper deals with the problem of safety verification of nonlinear hybrid systems. We start from a classical method that uses interval arithmetic to check whether trajectories can move over the boundaries in a rectangular grid. We put this method into an abstraction refinement framework and impr ..."
Abstract

Cited by 75 (11 self)
 Add to MetaCart
(Show Context)
This paper deals with the problem of safety verification of nonlinear hybrid systems. We start from a classical method that uses interval arithmetic to check whether trajectories can move over the boundaries in a rectangular grid. We put this method into an abstraction refinement framework and improve it by developing an additional refinement step that employs interval constraint propagation to add information to the abstraction without introducing new grid elements. Moreover, the resulting method allows switching conditions, initial states and unsafe states to be described by complex constraints instead of sets that correspond to grid elements. Nevertheless, the method can be easily implemented since it is based on a welldefined set of constraints, on which one can run any constraint propagation based solver. Tests of such an implementation are promising.
Interval arithmetic yields efficient dynamic filters for computational geometry
 Disc. Appl. Maths
"... We discuss floatingpoint filters as a means of restricting the precision needed for arithmetic operations while still computing the exact result. We show that interval techniques can be used to speed up the exact evaluation of geometric predicates and describe an efficient implementation of interva ..."
Abstract

Cited by 66 (13 self)
 Add to MetaCart
We discuss floatingpoint filters as a means of restricting the precision needed for arithmetic operations while still computing the exact result. We show that interval techniques can be used to speed up the exact evaluation of geometric predicates and describe an efficient implementation of interval arithmetic that is strongly influenced by the rounding modes of the widely used IEEE 754 standard. Using this approach we engineer an efficient floatingpoint filter for the computation of the sign of a determinant that works for arbitrary dimensions. We validate our approach experimentally, comparing it with other static, dynamic and semistatic filters. 1
Subdivision Direction Selection In Interval Methods For Global Optimization
 SIAM J. Numer. Anal
, 1997
"... . The role of the interval subdivision selection rule is investigated in branchandbound algorithms for global optimization. The class of rules that allow convergence for the model algorithm is characterized, and it is shown that the four rules investigated satisfy the conditions of convergence. A ..."
Abstract

Cited by 60 (19 self)
 Add to MetaCart
(Show Context)
. The role of the interval subdivision selection rule is investigated in branchandbound algorithms for global optimization. The class of rules that allow convergence for the model algorithm is characterized, and it is shown that the four rules investigated satisfy the conditions of convergence. A numerical study with a wide spectrum of test problems indicates that there are substantial differences between the rules in terms of the required CPU time, the number of function and derivative evaluations and space complexity, and two rules can provide substantial improvements in efficiency. Key words. global optimization, interval arithmetic, interval subdivision AMS subject classifications. 65K05, 90C30 Abbreviated title: Subdivision directions in interval methods. 1. Introduction. Interval subdivision methods for global optimization [7, 21] aim at providing reliable solutions to global optimization problems min x2X f(x) (1) where the objective function f : IR n ! IR is continuo...
Molecular Modeling Of Proteins And Mathematical Prediction Of Protein Structure
 SIAM Review
, 1997
"... . This paper discusses the mathematical formulation of and solution attempts for the socalled protein folding problem. The static aspect is concerned with how to predict the folded (native, tertiary) structure of a protein, given its sequence of amino acids. The dynamic aspect asks about the possib ..."
Abstract

Cited by 59 (5 self)
 Add to MetaCart
(Show Context)
. This paper discusses the mathematical formulation of and solution attempts for the socalled protein folding problem. The static aspect is concerned with how to predict the folded (native, tertiary) structure of a protein, given its sequence of amino acids. The dynamic aspect asks about the possible pathways to folding and unfolding, including the stability of the folded protein. From a mathematical point of view, there are several main sides to the static problem:  the selection of an appropriate potential energy function;  the parameter identification by fitting to experimental data; and  the global optimization of the potential. The dynamic problem entails, in addition, the solution of (because of multiple time scales very stiff) ordinary or stochastic differential equations (molecular dynamics simulation), or (in case of constrained molecular dynamics) of differentialalgebraic equations. A theme connecting the static and dynamic aspect is the determination and formation of...
A Review of Preconditioners for the Interval GaussSeidel Method
, 1991
"... . Interval Newton methods in conjunction with generalized bisection can form the basis of algorithms that find all real roots within a specified box X ae R n of a system of nonlinear equations F (X) = 0 with mathematical certainty, even in finiteprecision arithmetic. In such methods, the system ..."
Abstract

Cited by 58 (18 self)
 Add to MetaCart
. Interval Newton methods in conjunction with generalized bisection can form the basis of algorithms that find all real roots within a specified box X ae R n of a system of nonlinear equations F (X) = 0 with mathematical certainty, even in finiteprecision arithmetic. In such methods, the system F (X) = 0 is transformed into a linear interval system 0 = F (M) +F 0 (X)( ~ X \Gamma M); if interval arithmetic is then used to bound the solutions of this system, the resulting box ~ X contains all roots of the nonlinear system. We may use the interval GaussSeidel method to find these solution bounds. In order to increase the overall efficiency of the interval Newton / generalized bisection algorithm, the linear interval system is multiplied by a preconditioner matrix Y before the interval GaussSeidel method is applied. Here, we review results we have obtained over the past few years concerning computation of such preconditioners. We emphasize importance and connecting relationships,...
Interval Computations: Introduction, Uses, and Resources
 Euromath Bulletin
, 1996
"... ..."
(Show Context)