Results 1  10
of
60
On Projection Algorithms for Solving Convex Feasibility Problems
, 1996
"... Due to their extraordinary utility and broad applicability in many areas of classical mathematics and modern physical sciences (most notably, computerized tomography), algorithms for solving convex feasibility problems continue to receive great attention. To unify, generalize, and review some of the ..."
Abstract

Cited by 330 (44 self)
 Add to MetaCart
(Show Context)
Due to their extraordinary utility and broad applicability in many areas of classical mathematics and modern physical sciences (most notably, computerized tomography), algorithms for solving convex feasibility problems continue to receive great attention. To unify, generalize, and review some of these algorithms, a very broad and flexible framework is investigated . Several crucial new concepts which allow a systematic discussion of questions on behaviour in general Hilbert spaces and on the quality of convergence are brought out. Numerous examples are given. 1991 M.R. Subject Classification. Primary 47H09, 49M45, 6502, 65J05, 90C25; Secondary 26B25, 41A65, 46C99, 46N10, 47N10, 52A05, 52A41, 65F10, 65K05, 90C90, 92C55. Key words and phrases. Angle between two subspaces, averaged mapping, Cimmino's method, computerized tomography, convex feasibility problem, convex function, convex inequalities, convex programming, convex set, Fej'er monotone sequence, firmly nonexpansive mapping, H...
Incremental Subgradient Methods For Nondifferentiable Optimization
, 2001
"... We consider a class of subgradient methods for minimizing a convex function that consists of the sum of a large number of component functions. This type of minimization arises in a dual context from Lagrangian relaxation of the coupling constraints of large scale separable problems. The idea is to p ..."
Abstract

Cited by 118 (10 self)
 Add to MetaCart
We consider a class of subgradient methods for minimizing a convex function that consists of the sum of a large number of component functions. This type of minimization arises in a dual context from Lagrangian relaxation of the coupling constraints of large scale separable problems. The idea is to perform the subgradient iteration incrementally, by sequentially taking steps along the subgradients of the component functions, with intermediate adjustment of the variables after processing each component function. This incremental approach has been very successful in solving large differentiable least squares problems, such as those arising in the training of neural networks, and it has resulted in a much better practical rate of convergence than the steepest descent method. In this paper, we establish the convergence properties of a number of variants of incremental subgradient methods, including some that are stochastic. Based on the analysis and computational experiments, the methods appear very promising and effective for important classes of large problems. A particularly interesting discovery is that by randomizing the order of selection of component functions for iteration, the convergence rate is substantially improved.
A WEAKTOSTRONGCONVERGENCE PRINCIPLE FOR FEJÉRMONOTONE METHODS IN HILBERT SPACES
, 2001
"... We consider a wide class of iterative methods arising in numerical mathematics and optimization that are known to converge only weakly. Exploiting an idea originally proposed by Haugazeau, we present a simple modification of these methods that makes them strongly convergent without additional assump ..."
Abstract

Cited by 81 (13 self)
 Add to MetaCart
We consider a wide class of iterative methods arising in numerical mathematics and optimization that are known to converge only weakly. Exploiting an idea originally proposed by Haugazeau, we present a simple modification of these methods that makes them strongly convergent without additional assumptions. Several applications are discussed.
An Adaptive Level Set Method for Nondifferentiable Constrained Image Recovery
 IEEE TRANS. IMAGE PROCESSING
, 2002
"... The formulation of a wide variety of image recovery problems leads to the minimization of a convex objective over a convex set representing the constraints derived from a priori knowledge and consistency with the observed signals. In recent years, nondifferentiable objectives have become popular due ..."
Abstract

Cited by 36 (4 self)
 Add to MetaCart
The formulation of a wide variety of image recovery problems leads to the minimization of a convex objective over a convex set representing the constraints derived from a priori knowledge and consistency with the observed signals. In recent years, nondifferentiable objectives have become popular due in part to their ability to capture certain features such as sharp edges. They also arise naturally in minimax inconsistent set theoretic recovery problems. At the same time, the issue of developing reliable numerical algorithms to solve such convex programs in the context of image recovery applications has received little attention. In this paper, we address this issue and propose an adaptive level set method for nondifferentiable constrained image recovery. The asymptotic properties of the method are analyzed and its implementation is discussed. Numerical experiments illustrate applications to total variation and minimax set theoretic image restoration and denoising problems.
Scheduling Of Manufacturing Systems Using The Lagrangian Relaxation Technique
 IEEE Transactions on Automatic Control
, 1993
"... Scheduling is one of the most basic but the most difficult problems encountered in the manufacturing industry. Generally, some degree of timeconsuming and impractical enumeration is required to obtain optimal solutions. Industry has thus relied on a combination of heuristics and simulation to solve ..."
Abstract

Cited by 34 (9 self)
 Add to MetaCart
Scheduling is one of the most basic but the most difficult problems encountered in the manufacturing industry. Generally, some degree of timeconsuming and impractical enumeration is required to obtain optimal solutions. Industry has thus relied on a combination of heuristics and simulation to solve the problem, resulting in unreliable and often infeasible schedules. Yet, there is a great need for an improvement in scheduling operations in complex and turbulent manufacturing environments. The logical strategy is to find scheduling methods which consistently generate good schedules efficiently. However, it is often difficult to measure the quality of a schedule without knowing the optimum. In this paper, the practical scheduling of three manufacturing environments are examined in the increasing order of complexity. The first problem considers scheduling singleoperation jobs on parallel, identical machines; the second one is concerned with scheduling multipleoperation jobs with simple ...
DUAL COORDINATE STEP METHODS FOR LINEAR NETWORK FLOW PROBLEMS
, 1988
"... We review a class of recentlyproposed linearcost network flow methods which are amenable to distributed implementation. All the methods in the class use the notion of ecomplementary slackness, and most do not explicitly manipulate any "global " objects such as paths, trees, or cuts. Int ..."
Abstract

Cited by 31 (8 self)
 Add to MetaCart
We review a class of recentlyproposed linearcost network flow methods which are amenable to distributed implementation. All the methods in the class use the notion of ecomplementary slackness, and most do not explicitly manipulate any "global " objects such as paths, trees, or cuts. Interestingly, these methods have stimulated a large number of new serial computational complexity results. We develop the basic theory of these methods and present two specific methods, the erelaxation algorithm for the minimumcost flow problem, and the auction algorithm for the assignment problem. We show how to implement these methods with serial complexities of O(N 3 log NC) and O(NA log NC), respectively. We also discuss practical implementation issues and computational experience to date. Finally, we show how to implement erelaxation in a completely asynchronous, "chaotic" environment in which some processors compute faster than others, some processors communicate faster than others, and there can be arbitrarily large communication delays.
Ergodic, primal convergence in dual subgradient schemes for convex programming
 TO APPEAR IN MATHEMATICAL PROGAMMING
"... Lagrangean dualization and subgradient optimization techniques are frequently used within the field of computational optimization for nding approximate solutions to large, structured optimization problems. The dual subgradient scheme does not automatically produce primal feasible solutions; there i ..."
Abstract

Cited by 25 (2 self)
 Add to MetaCart
Lagrangean dualization and subgradient optimization techniques are frequently used within the field of computational optimization for nding approximate solutions to large, structured optimization problems. The dual subgradient scheme does not automatically produce primal feasible solutions; there is an abundance of techniques for computing such solutions (via penalty functions, tangential approximation schemes, or the solution of auxiliary primal programs), all of which require a fair amount of computational effort. We consider a subgradient optimization scheme applied to a Lagrangean dual formulation of a convex program, and construct, at minor cost, an ergodic sequence of subproblem solutions which converges to the primal solution set. Numerical experiments performed on a trac equilibrium assignment problem under road pricing show that the computation of the ergodic sequence results in a considerable improvement in the quality of the primal solutions obtained, compared to those generated in the basic subgradient scheme.
Extrapolation algorithm for affineconvex feasibility problems
, 2005
"... The convex feasibility problem under consideration is to find a common point of a countable family of closed affine subspaces and convex sets in a Hilbert space. To solve such problems, we propose a general parallel blockiterative algorithmic framework in which the affine subspaces are exploited to ..."
Abstract

Cited by 24 (6 self)
 Add to MetaCart
The convex feasibility problem under consideration is to find a common point of a countable family of closed affine subspaces and convex sets in a Hilbert space. To solve such problems, we propose a general parallel blockiterative algorithmic framework in which the affine subspaces are exploited to introduce extrapolated overrelaxations. This framework encompasses a wide range of projection, subgradient projection, proximal, and fixed point methods encountered in various branches of applied mathematics. The asymptotic behavior of the method is investigated and numerical experiments are provided to illustrate the benefits of the extrapolations.
Conditional Subgradient Optimization  Theory and Applications
 European Journal of Operational Research
, 1996
"... We generalize the subgradient optimization method for nondifferentiable convex programming to utilize conditional subgradients. Firstly, we derive the new method and establish its convergence by generalizing convergence results for traditional subgradient optimization. Secondly, we consider a partic ..."
Abstract

Cited by 17 (4 self)
 Add to MetaCart
We generalize the subgradient optimization method for nondifferentiable convex programming to utilize conditional subgradients. Firstly, we derive the new method and establish its convergence by generalizing convergence results for traditional subgradient optimization. Secondly, we consider a particular choice of conditional subgradients, obtained by projections, which leads to an easily implementable modification of traditional subgradient optimization schemes. To evaluate the subgradient projection method we consider its use in three applications: uncapacitated facility location, twoperson zerosum matrix games, and multicommodity network flows. Computational experiments show that the subgradient projection method performs better than traditional subgradient optimization; in some cases the difference is considerable. These results suggest that our simple modification may improve subgradient optimization schemes significantly. This finding is important as such schemes are very popula...