Results 1  10
of
3,591,481
Onevariable word equations in linear time
, 2013
"... Abstract. In this paper we consider word equations with one variable (and arbitrary many occurrences of it). A recent technique of recompression, which is applicable to general word equations, is shown to be suitable also in this case. While in general case the recompression is nondeterministic it ..."
Abstract

Cited by 4 (2 self)
 Add to MetaCart
Abstract. In this paper we consider word equations with one variable (and arbitrary many occurrences of it). A recent technique of recompression, which is applicable to general word equations, is shown to be suitable also in this case. While in general case the recompression is non
Lambertian Reflectance and Linear Subspaces
, 2000
"... We prove that the set of all reflectance functions (the mapping from surface normals to intensities) produced by Lambertian objects under distant, isotropic lighting lies close to a 9D linear subspace. This implies that, in general, the set of images of a convex Lambertian object obtained under a wi ..."
Abstract

Cited by 514 (20 self)
 Add to MetaCart
We prove that the set of all reflectance functions (the mapping from surface normals to intensities) produced by Lambertian objects under distant, isotropic lighting lies close to a 9D linear subspace. This implies that, in general, the set of images of a convex Lambertian object obtained under a
For Most Large Underdetermined Systems of Linear Equations the Minimal ℓ1norm Solution is also the Sparsest Solution
 Comm. Pure Appl. Math
, 2004
"... We consider linear equations y = Φα where y is a given vector in R n, Φ is a given n by m matrix with n < m ≤ An, and we wish to solve for α ∈ R m. We suppose that the columns of Φ are normalized to unit ℓ 2 norm 1 and we place uniform measure on such Φ. We prove the existence of ρ = ρ(A) so that ..."
Abstract

Cited by 560 (10 self)
 Add to MetaCart
We consider linear equations y = Φα where y is a given vector in R n, Φ is a given n by m matrix with n < m ≤ An, and we wish to solve for α ∈ R m. We suppose that the columns of Φ are normalized to unit ℓ 2 norm 1 and we place uniform measure on such Φ. We prove the existence of ρ = ρ(A) so
LSQR: An Algorithm for Sparse Linear Equations and Sparse Least Squares
 ACM Trans. Math. Software
, 1982
"... An iterative method is given for solving Ax ~ffi b and minU Ax b 112, where the matrix A is large and sparse. The method is based on the bidiagonalization procedure of Golub and Kahan. It is analytically equivalent to the standard method of conjugate gradients, but possesses more favorable numerica ..."
Abstract

Cited by 649 (21 self)
 Add to MetaCart
gradient algorithms, indicating that I~QR is the most reliable algorithm when A is illconditioned. Categories and Subject Descriptors: G.1.2 [Numerical Analysis]: ApprorJmationleast squares approximation; G.1.3 [Numerical Analysis]: Numerical Linear Algebralinear systems (direct and
OneVariable Models
"... by any electronic or mechanical means (including photocopying, recording, or information storage and retrieval) without permission in writing from the publisher, except for reading and browsing via the World Wide Web. Users are not permitted to mount this file on any network servers. Follow links Cl ..."
Abstract

Cited by 2 (0 self)
 Add to MetaCart
by any electronic or mechanical means (including photocopying, recording, or information storage and retrieval) without permission in writing from the publisher, except for reading and browsing via the World Wide Web. Users are not permitted to mount this file on any network servers. Follow links Class Use and other Permissions. For more information, send email to:
Using Linear Algebra for Intelligent Information Retrieval
 SIAM REVIEW
, 1995
"... Currently, most approaches to retrieving textual materials from scientific databases depend on a lexical match between words in users' requests and those in or assigned to documents in a database. Because of the tremendous diversity in the words people use to describe the same document, lexical ..."
Abstract

Cited by 672 (18 self)
 Add to MetaCart
Currently, most approaches to retrieving textual materials from scientific databases depend on a lexical match between words in users' requests and those in or assigned to documents in a database. Because of the tremendous diversity in the words people use to describe the same document
Understanding Normal and Impaired Word Reading: Computational Principles in QuasiRegular Domains
 PSYCHOLOGICAL REVIEW
, 1996
"... We develop a connectionist approach to processing in quasiregular domains, as exemplified by English word reading. A consideration of the shortcomings of a previous implementation (Seidenberg & McClelland, 1989, Psych. Rev.) in reading nonwords leads to the development of orthographic and phono ..."
Abstract

Cited by 583 (94 self)
 Add to MetaCart
We develop a connectionist approach to processing in quasiregular domains, as exemplified by English word reading. A consideration of the shortcomings of a previous implementation (Seidenberg & McClelland, 1989, Psych. Rev.) in reading nonwords leads to the development of orthographic
Why a diagram is (sometimes) worth ten thousand words
 Cognitive Science
, 1987
"... We distinguish diagrammatic from sentential paperandpencil representationsof information by developing alternative models of informationprocessing systems that are informationally equivalent and that can be characterized as sentential or diagrammatic. Sentential representations are sequential, li ..."
Abstract

Cited by 777 (2 self)
 Add to MetaCart
We distinguish diagrammatic from sentential paperandpencil representationsof information by developing alternative models of informationprocessing systems that are informationally equivalent and that can be characterized as sentential or diagrammatic. Sentential representations are sequential, like the propositions in a text. Dlogrammotlc representations ore indexed by location in a plane. Diogrommatic representations also typically display information that is only implicit in sententiol representations and that therefore has to be computed, sometimes at great cost, to make it explicit for use. We then contrast the computational efficiency of these representotions for solving several illustrative problems in mothematics and physics. When two representotions are informationally equivolent, their computational efficiency depends on the informationprocessing operators that act on them. Two sets of operators may differ in their copobilities for recognizing patterns, in the inferences they con carry out directly, and in their control strategies (in portitular. the control of search). Diogrommotic ond sentential representations sup
Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences
 ACOUSTICS, SPEECH AND SIGNAL PROCESSING, IEEE TRANSACTIONS ON
, 1980
"... Several parametric representations of the acoustic signal were compared as to word recognition performance in a syllableoriented continuous speech recognition system. The vocabulary included many phonetically similar monosyllabic words, therefore the emphasis was on ability to retain phonetically ..."
Abstract

Cited by 1089 (2 self)
 Add to MetaCart
phonetically significant acoustic information in the face of syntactic and duration variations. For each ~ arameter set (based on a melfrequency cepstrum, a linear frequency cepstrum, a linear prediction cepstrum, a linear prediction spectrum, or a set of reflection coefficients), word templates were
WordNet: An online lexical database
 International Journal of Lexicography
, 1990
"... WordNet is an online lexical reference system whose design is inspired by current ..."
Abstract

Cited by 1945 (9 self)
 Add to MetaCart
WordNet is an online lexical reference system whose design is inspired by current
Results 1  10
of
3,591,481