Results 1  10
of
98,698
Unified analysis of discontinuous Galerkin methods for elliptic problems
 SIAM J. Numer. Anal
, 2001
"... Abstract. We provide a framework for the analysis of a large class of discontinuous methods for secondorder elliptic problems. It allows for the understanding and comparison of most of the discontinuous Galerkin methods that have been proposed over the past three decades for the numerical treatment ..."
Abstract

Cited by 519 (31 self)
 Add to MetaCart
Abstract. We provide a framework for the analysis of a large class of discontinuous methods for secondorder elliptic problems. It allows for the understanding and comparison of most of the discontinuous Galerkin methods that have been proposed over the past three decades for the numerical
HOMOGENIZATION AND TWOSCALE CONVERGENCE
, 1992
"... Following an idea of G. Nguetseng, the author defines a notion of "twoscale" convergence, which is aimed at a better description of sequences of oscillating functions. Bounded sequences in L2(f) are proven to be relatively compact with respect to this new type of convergence. A corrector ..."
Abstract

Cited by 445 (14 self)
 Add to MetaCart
linear and nonlinear secondorder elliptic equations.
Symmetry and Related Properties via the Maximum Principle
, 1979
"... We prove symmetry, and some related properties, of positive solutions of second order elliptic equations. Our methods employ various forms of the maximum principle, and a device of moving parallel planes to a critical position, and then showing that the solution is symmetric about the limiting plan ..."
Abstract

Cited by 539 (4 self)
 Add to MetaCart
We prove symmetry, and some related properties, of positive solutions of second order elliptic equations. Our methods employ various forms of the maximum principle, and a device of moving parallel planes to a critical position, and then showing that the solution is symmetric about the limiting
Region Competition: Unifying Snakes, Region Growing, and Bayes/MDL for Multiband Image Segmentation
 IEEE Transactions on Pattern Analysis and Machine Intelligence
, 1996
"... We present a novel statistical and variational approach to image segmentation based on a new algorithm named region competition. This algorithm is derived by minimizing a generalized Bayes/MDL criterion using the variational principle. The algorithm is guaranteed to converge to a local minimum and c ..."
Abstract

Cited by 778 (21 self)
 Add to MetaCart
We present a novel statistical and variational approach to image segmentation based on a new algorithm named region competition. This algorithm is derived by minimizing a generalized Bayes/MDL criterion using the variational principle. The algorithm is guaranteed to converge to a local minimum
DecisionTheoretic Planning: Structural Assumptions and Computational Leverage
 JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH
, 1999
"... Planning under uncertainty is a central problem in the study of automated sequential decision making, and has been addressed by researchers in many different fields, including AI planning, decision analysis, operations research, control theory and economics. While the assumptions and perspectives ..."
Abstract

Cited by 510 (4 self)
 Add to MetaCart
or plans. Planning problems commonly possess structure in the reward and value functions used to de...
Convex Analysis
, 1970
"... In this book we aim to present, in a unified framework, a broad spectrum of mathematical theory that has grown in connection with the study of problems of optimization, equilibrium, control, and stability of linear and nonlinear systems. The title Variational Analysis reflects this breadth. For a lo ..."
Abstract

Cited by 5350 (67 self)
 Add to MetaCart
In this book we aim to present, in a unified framework, a broad spectrum of mathematical theory that has grown in connection with the study of problems of optimization, equilibrium, control, and stability of linear and nonlinear systems. The title Variational Analysis reflects this breadth. For a
Least angle regression
 Ann. Statist
"... The purpose of model selection algorithms such as All Subsets, Forward Selection and Backward Elimination is to choose a linear model on the basis of the same set of data to which the model will be applied. Typically we have available a large collection of possible covariates from which we hope to s ..."
Abstract

Cited by 1308 (43 self)
 Add to MetaCart
implements the Lasso, an attractive version of ordinary least squares that constrains the sum of the absolute regression coefficients; the LARS modification calculates all possible Lasso estimates for a given problem, using an order of magnitude less computer time than previous methods. (2) A different LARS
A Learning Algorithm for Continually Running Fully Recurrent Neural Networks
, 1989
"... The exact form of a gradientfollowing learning algorithm for completely recurrent networks running in continually sampled time is derived and used as the basis for practical algorithms for temporal supervised learning tasks. These algorithms have: (1) the advantage that they do not require a precis ..."
Abstract

Cited by 529 (4 self)
 Add to MetaCart
the retention of information over time periods having either fixed or indefinite length. 1 Introduction A major problem in connectionist theory is to develop learning algorithms that can tap the full computational power of neural networks. Much progress has been made with feedforward networks, and attention
Numerical integration of the Cartesian equations of motion of a system with constraints: molecular dynamics of nalkanes
 J. Comput. Phys
, 1977
"... A numerical algorithm integrating the 3N Cartesian equations of motion of a system of N points subject to holonomic constraints is formulated. The relations of constraint remain perfectly fulfilled at each step of the trajectory despite the approximate character of numerical integration. The method ..."
Abstract

Cited by 682 (6 self)
 Add to MetaCart
(MD), which has been widely used in the past for studying simple liquids and solids, has more recently been applied to molecular systems with internal degrees of freedom such as N, [l], H,O [2] and even C,H,, [3]. In applying the MD method three problems arise: (a) the choice of a suitable mechanical
Regression Shrinkage and Selection Via the Lasso
 Journal of the Royal Statistical Society, Series B
, 1994
"... We propose a new method for estimation in linear models. The "lasso" minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant. Because of the nature of this constraint it tends to produce some coefficients that are exactl ..."
Abstract

Cited by 4055 (51 self)
 Add to MetaCart
We propose a new method for estimation in linear models. The "lasso" minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant. Because of the nature of this constraint it tends to produce some coefficients
Results 1  10
of
98,698