Results 1  10
of
846,095
Submodule Construction as Equation Solving in CCS
 IN PROCEEDINGS OF THE FOUNDATIONS OF SOFTWARE TECHNOLOGY AND THEORETICAL COMPUTER SCIENCE, LECTURE NOTES IN COMPUTER SCIENCE 287
, 1989
"... In topdown design methodologies the following problem arises: given specifications of a system and of some of its submodules, derive a specification for the remaining submodules. We formulate this problem in CCS as an equation (AjX)nL B, where X is unknown, B represents the whole system, A the ..."
Abstract

Cited by 6 (2 self)
 Add to MetaCart
In topdown design methodologies the following problem arises: given specifications of a system and of some of its submodules, derive a specification for the remaining submodules. We formulate this problem in CCS as an equation (AjX)nL B, where X is unknown, B represents the whole system, A
Cognitive load during problem solving: effects on learning
 COGNITIVE SCIENCE
, 1988
"... Considerable evidence indicates that domain specific knowledge in the form of schemes is the primary factor distinguishing experts from novices in problemsolving skill. Evidence that conventional problemsolving activity is not effective in schema acquisition is also accumulating. It is suggested t ..."
Abstract

Cited by 603 (13 self)
 Add to MetaCart
Considerable evidence indicates that domain specific knowledge in the form of schemes is the primary factor distinguishing experts from novices in problemsolving skill. Evidence that conventional problemsolving activity is not effective in schema acquisition is also accumulating. It is suggested
GMRES: A generalized minimal residual algorithm for solving nonsymmetric linear systems
 SIAM J. SCI. STAT. COMPUT
, 1986
"... We present an iterative method for solving linear systems, which has the property ofminimizing at every step the norm of the residual vector over a Krylov subspace. The algorithm is derived from the Arnoldi process for constructing an l2orthogonal basis of Krylov subspaces. It can be considered a ..."
Abstract

Cited by 2046 (40 self)
 Add to MetaCart
We present an iterative method for solving linear systems, which has the property ofminimizing at every step the norm of the residual vector over a Krylov subspace. The algorithm is derived from the Arnoldi process for constructing an l2orthogonal basis of Krylov subspaces. It can be considered
The Lifting Scheme: A Construction Of Second Generation Wavelets
, 1997
"... . We present the lifting scheme, a simple construction of second generation wavelets, wavelets that are not necessarily translates and dilates of one fixed function. Such wavelets can be adapted to intervals, domains, surfaces, weights, and irregular samples. We show how the lifting scheme leads to ..."
Abstract

Cited by 541 (16 self)
 Add to MetaCart
. We present the lifting scheme, a simple construction of second generation wavelets, wavelets that are not necessarily translates and dilates of one fixed function. Such wavelets can be adapted to intervals, domains, surfaces, weights, and irregular samples. We show how the lifting scheme leads
The selfduality equations on a Riemann surface
 Proc. Lond. Math. Soc., III. Ser
, 1987
"... In this paper we shall study a special class of solutions of the selfdual YangMills equations. The original selfduality equations which arose in mathematical physics were defined on Euclidean 4space. The physically relevant solutions were the ones with finite action—the socalled 'instanton ..."
Abstract

Cited by 524 (6 self)
 Add to MetaCart
In this paper we shall study a special class of solutions of the selfdual YangMills equations. The original selfduality equations which arose in mathematical physics were defined on Euclidean 4space. The physically relevant solutions were the ones with finite action—the socalled &apos
LSQR: An Algorithm for Sparse Linear Equations and Sparse Least Squares
 ACM Trans. Math. Software
, 1982
"... An iterative method is given for solving Ax ~ffi b and minU Ax b 112, where the matrix A is large and sparse. The method is based on the bidiagonalization procedure of Golub and Kahan. It is analytically equivalent to the standard method of conjugate gradients, but possesses more favorable numerica ..."
Abstract

Cited by 649 (21 self)
 Add to MetaCart
An iterative method is given for solving Ax ~ffi b and minU Ax b 112, where the matrix A is large and sparse. The method is based on the bidiagonalization procedure of Golub and Kahan. It is analytically equivalent to the standard method of conjugate gradients, but possesses more favorable
Lag length selection and the construction of unit root tests with good size and power
 Econometrica
, 2001
"... It is widely known that when there are errors with a movingaverage root close to −1, a high order augmented autoregression is necessary for unit root tests to have good size, but that information criteria such as the AIC and the BIC tend to select a truncation lag (k) that is very small. We conside ..."
Abstract

Cited by 534 (14 self)
 Add to MetaCart
It is widely known that when there are errors with a movingaverage root close to −1, a high order augmented autoregression is necessary for unit root tests to have good size, but that information criteria such as the AIC and the BIC tend to select a truncation lag (k) that is very small. We consider a class of Modified Information Criteria (MIC) with a penalty factor that is sample dependent. It takes into account the fact that the bias in the sum of the autoregressive coefficients is highly dependent on k and adapts to the type of deterministic components present. We use a local asymptotic framework in which the movingaverage root is local to −1 to document how the MIC performs better in selecting appropriate values of k. In montecarlo experiments, the MIC is found to yield huge size improvements to the DF GLS and the feasible point optimal PT test developed in Elliott, Rothenberg and Stock (1996). We also extend the M tests developed in Perron and Ng (1996) to allow for GLS detrending of the data. The MIC along with GLS detrended data yield a set of tests with desirable size and power properties.
For Most Large Underdetermined Systems of Linear Equations the Minimal ℓ1norm Solution is also the Sparsest Solution
 Comm. Pure Appl. Math
, 2004
"... We consider linear equations y = Φα where y is a given vector in R n, Φ is a given n by m matrix with n < m ≤ An, and we wish to solve for α ∈ R m. We suppose that the columns of Φ are normalized to unit ℓ 2 norm 1 and we place uniform measure on such Φ. We prove the existence of ρ = ρ(A) so that ..."
Abstract

Cited by 560 (10 self)
 Add to MetaCart
We consider linear equations y = Φα where y is a given vector in R n, Φ is a given n by m matrix with n < m ≤ An, and we wish to solve for α ∈ R m. We suppose that the columns of Φ are normalized to unit ℓ 2 norm 1 and we place uniform measure on such Φ. We prove the existence of ρ = ρ(A) so
Asymptotic Confidence Intervals for Indirect Effects in Structural EQUATION MODELS
 IN SOCIOLOGICAL METHODOLOGY
, 1982
"... ..."
Results 1  10
of
846,095