Results 1  10
of
385,238
Theoretical improvements in algorithmic efficiency for network flow problems

, 1972
"... This paper presents new algorithms for the maximum flow problem, the Hitchcock transportation problem, and the general minimumcost flow problem. Upper bounds on ... the numbers of steps in these algorithms are derived, and are shown to compale favorably with upper bounds on the numbers of steps req ..."
Abstract

Cited by 565 (0 self)
 Add to MetaCart
are given. We show that, if each flow augmentation is made along an augmenting path having a minimum number of arcs, then a maximum flow in an nnode network will be obtained after no more than ~(n a n) augmentations; and then we show that if each flow change is chosen to produce a maximum increase
VERY HIGH RESOLUTION INTERPOLATED CLIMATE SURFACES FOR GLOBAL LAND AREAS
, 2005
"... We developed interpolated climate surfaces for global land areas (excluding Antarctica) at a spatial resolution of 30 arc s (often referred to as 1km spatial resolution). The climate elements considered were monthly precipitation and mean, minimum, and maximum temperature. Input data were gathered ..."
Abstract

Cited by 499 (7 self)
 Add to MetaCart
We developed interpolated climate surfaces for global land areas (excluding Antarctica) at a spatial resolution of 30 arc s (often referred to as 1km spatial resolution). The climate elements considered were monthly precipitation and mean, minimum, and maximum temperature. Input data were gathered
An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants
 MACHINE LEARNING
, 1999
"... Methods for voting classification algorithms, such as Bagging and AdaBoost, have been shown to be very successful in improving the accuracy of certain classifiers for artificial and realworld datasets. We review these algorithms and describe a large empirical study comparing several variants in co ..."
Abstract

Cited by 695 (2 self)
 Add to MetaCart
decomposition of the error to show how different
methods and variants influence these two terms. This allowed us to
determine that Bagging reduced variance of unstable methods, while
boosting methods (AdaBoost and Arcx4) reduced both the bias and
variance of unstable methods but increased the variance
Halfa century of research on the Stroop effect: An integrative review
 PsychologicalBulletin
, 1991
"... The literature on interference in the Stroop ColorWord Task, covering over 50 years and some 400 studies, is organized and reviewed. In so doing, a set ofl 8 reliable empirical findings is isolated that must be captured by any successful theory of the Stroop effect. Existing theoretical positions a ..."
Abstract

Cited by 621 (14 self)
 Add to MetaCart
dimensions are likely to be more successful than are earlier theories attempting to locate a single bottleneck in attention. In 1935, J. R. Stroop published his landmark article on attention and interference, an article more influential now than it was then. Why has the Stroop task continued to fascinate us
Static Scheduling of Synchronous Data Flow Programs for Digital Signal Processing
 IEEE TRANSACTIONS ON COMPUTERS
, 1987
"... Large grain data flow (LGDF) programming is natural and convenient for describing digital signal processing (DSP) systems, but its runtime overhead is costly in real time or costsensitive applications. In some situations, designers are not willing to squander computing resources for the sake of pro ..."
Abstract

Cited by 592 (37 self)
 Add to MetaCart
Large grain data flow (LGDF) programming is natural and convenient for describing digital signal processing (DSP) systems, but its runtime overhead is costly in real time or costsensitive applications. In some situations, designers are not willing to squander computing resources for the sake of programmer convenience. This is particularly true when the target machine is a programmable DSP chip. However, the runtime overhead inherent in most LGDF implementations is not required for most signal processing systems because such systems are mostly synchronous (in the DSP sense). Synchronous data flow (SDF) differs from traditional data flow in that the amount of data produced and consumed by a data flow node is specified a priori for each input and output. This is equivalent to specifying the relative sample rates in signal processing system. This means that the scheduling of SDF nodes need not be done at runtime, but can be done at compile time (statically), so the runtime overhead evaporates. The sample rates can all be different, which is not true of most current datadriven digital signal processing programming methodologies. Synchronous data flow is closely related to computation graphs, a special case of Petri nets. This selfcontained paper develops the theory necessary to statically schedule SDF programs on single or multiple processors. A class of static (compile time) scheduling algorithms is proven valid, and specific algorithms are given for scheduling SDF systems onto single or multiple processors.
Automatic Discovery of Linear Restraints Among Variables of a Program
, 1978
"... The model of abstract interpretation of programs developed by Cousot and Cousot [2nd ISOP, 1976], Cousot and Cousot [POPL 1977] and Cousot [PhD thesis 1978] is applied to the static determination of linear equality or inequality invariant relations among numerical variables of programs. ..."
Abstract

Cited by 733 (47 self)
 Add to MetaCart
The model of abstract interpretation of programs developed by Cousot and Cousot [2nd ISOP, 1976], Cousot and Cousot [POPL 1977] and Cousot [PhD thesis 1978] is applied to the static determination of linear equality or inequality invariant relations among numerical variables of programs.
Constraint Networks
, 1992
"... Constraintbased reasoning is a paradigm for formulating knowledge as a set of constraints without specifying the method by which these constraints are to be satisfied. A variety of techniques have been developed for finding partial or complete solutions for different kinds of constraint expression ..."
Abstract

Cited by 1149 (43 self)
 Add to MetaCart
Constraintbased reasoning is a paradigm for formulating knowledge as a set of constraints without specifying the method by which these constraints are to be satisfied. A variety of techniques have been developed for finding partial or complete solutions for different kinds of constraint expressions. These have been successfully applied to diverse tasks such as design, diagnosis, truth maintenance, scheduling, spatiotemporal reasoning, logic programming and user interface. Constraint networks are graphical representations used to guide strategies for solving constraint satisfaction problems (CSPs).
The Valuation of Options for Alternative Stochastic Processes
 Journal of Financial Economics
, 1976
"... This paper examines the structure of option valuation problems and develops a new technique for their solution. It also introduces several jump and diffusion processes which have nol been used in previous models. The technique is applied lo these processes to find explicit option valuation formulas, ..."
Abstract

Cited by 661 (4 self)
 Add to MetaCart
This paper examines the structure of option valuation problems and develops a new technique for their solution. It also introduces several jump and diffusion processes which have nol been used in previous models. The technique is applied lo these processes to find explicit option valuation formulas, and solutions to some previously unsolved problems involving the pricing ofsecurities with payouts and potential bankruptcy. 1.
Region Competition: Unifying Snakes, Region Growing, and Bayes/MDL for Multiband Image Segmentation
 IEEE Transactions on Pattern Analysis and Machine Intelligence
, 1996
"... We present a novel statistical and variational approach to image segmentation based on a new algorithm named region competition. This algorithm is derived by minimizing a generalized Bayes/MDL criterion using the variational principle. The algorithm is guaranteed to converge to a local minimum and c ..."
Abstract

Cited by 778 (21 self)
 Add to MetaCart
We present a novel statistical and variational approach to image segmentation based on a new algorithm named region competition. This algorithm is derived by minimizing a generalized Bayes/MDL criterion using the variational principle. The algorithm is guaranteed to converge to a local minimum and combines aspects of snakes/balloons and region growing. Indeed the classic snakes/balloons and region growing algorithms can be directly derived from our approach. We provide theoretical analysis of region competition including accuracy of boundary location, criteria for initial conditions, and the relationship to edge detection using filters. It is straightforward to generalize the algorithm to multiband segmentation and we demonstrate it on grey level images, color images and texture images. The novel color model allows us to eliminate intensity gradients and shadows, thereby obtaining segmentation based on the albedos of objects. It also helps detect highlight regions. 1 Division of Appli...
Markov Random Field Models in Computer Vision
, 1994
"... . A variety of computer vision problems can be optimally posed as Bayesian labeling in which the solution of a problem is defined as the maximum a posteriori (MAP) probability estimate of the true labeling. The posterior probability is usually derived from a prior model and a likelihood model. The l ..."
Abstract

Cited by 515 (18 self)
 Add to MetaCart
. A variety of computer vision problems can be optimally posed as Bayesian labeling in which the solution of a problem is defined as the maximum a posteriori (MAP) probability estimate of the true labeling. The posterior probability is usually derived from a prior model and a likelihood model. The latter relates to how data is observed and is problem domain dependent. The former depends on how various prior constraints are expressed. Markov Random Field Models (MRF) theory is a tool to encode contextual constraints into the prior probability. This paper presents a unified approach for MRF modeling in low and high level computer vision. The unification is made possible due to a recent advance in MRF modeling for high level object recognition. Such unification provides a systematic approach for vision modeling based on sound mathematical principles. 1 Introduction Since its beginning in early 1960's, computer vision research has been evolving from heuristic design of algorithms to syste...
Results 1  10
of
385,238