Results 1  10
of
90,527
EIGENVALUES AND EXPANDERS
 COMBINATORICA
, 1986
"... Linear expanders have numerous applications to theoretical computer science. Here we show that a regular bipartite graph is an expander ifandonly if the second largest eigenvalue of its adjacency matrix is well separated from the first. This result, which has an analytic analogue for Riemannian mani ..."
Abstract

Cited by 404 (19 self)
 Add to MetaCart
Linear expanders have numerous applications to theoretical computer science. Here we show that a regular bipartite graph is an expander ifandonly if the second largest eigenvalue of its adjacency matrix is well separated from the first. This result, which has an analytic analogue for Riemannian
Eigenvalues and Expansion of Regular Graphs
 Journal of the ACM
, 1995
"... The spectral method is the best currently known technique to prove lower bounds on expansion. Ramanujan graphs, which have asymptotically optimal second eigenvalue, are the best known explicit expanders. The spectral method yielded a lower bound of k=4 on the expansion of linear sized subsets of kr ..."
Abstract

Cited by 50 (1 self)
 Add to MetaCart
The spectral method is the best currently known technique to prove lower bounds on expansion. Ramanujan graphs, which have asymptotically optimal second eigenvalue, are the best known explicit expanders. The spectral method yielded a lower bound of k=4 on the expansion of linear sized subsets of kregular
Nonlinear component analysis as a kernel eigenvalue problem

, 1996
"... We describe a new method for performing a nonlinear form of Principal Component Analysis. By the use of integral operator kernel functions, we can efficiently compute principal components in highdimensional feature spaces, related to input space by some nonlinear map; for instance the space of all ..."
Abstract

Cited by 1554 (85 self)
 Add to MetaCart
We describe a new method for performing a nonlinear form of Principal Component Analysis. By the use of integral operator kernel functions, we can efficiently compute principal components in highdimensional feature spaces, related to input space by some nonlinear map; for instance the space of all possible 5pixel products in 16x16 images. We give the derivation of the method, along with a discussion of other techniques which can be made nonlinear with the kernel approach; and present first experimental results on nonlinear feature extraction for pattern recognition.
Community detection in graphs
, 2009
"... The modern science of networks has brought significant advances to our understanding of complex systems. One of the most relevant features of graphs representing real systems is community structure, or clustering, i. e. the organization of vertices in clusters, with many edges joining vertices of th ..."
Abstract

Cited by 801 (1 self)
 Add to MetaCart
The modern science of networks has brought significant advances to our understanding of complex systems. One of the most relevant features of graphs representing real systems is community structure, or clustering, i. e. the organization of vertices in clusters, with many edges joining vertices
Inverse Acoustic and Electromagnetic Scattering Theory, Second Edition
, 1998
"... Abstract. This paper is a survey of the inverse scattering problem for timeharmonic acoustic and electromagnetic waves at fixed frequency. We begin by a discussion of “weak scattering ” and Newtontype methods for solving the inverse scattering problem for acoustic waves, including a brief discussi ..."
Abstract

Cited by 1072 (45 self)
 Add to MetaCart
discussion of Tikhonov’s method for the numerical solution of illposed problems. We then proceed to prove a uniqueness theorem for the inverse obstacle problems for acoustic waves and the linear sampling method for reconstructing the shape of a scattering obstacle from far field data. Included in our
The program dependence graph and its use in optimization
 ACM Transactions on Programming Languages and Systems
, 1987
"... In this paper we present an intermediate program representation, called the program dependence graph (PDG), that makes explicit both the data and control dependence5 for each operation in a program. Data dependences have been used to represent only the relevant data flow relationships of a program. ..."
Abstract

Cited by 989 (3 self)
 Add to MetaCart
In this paper we present an intermediate program representation, called the program dependence graph (PDG), that makes explicit both the data and control dependence5 for each operation in a program. Data dependences have been used to represent only the relevant data flow relationships of a program
Using Linear Algebra for Intelligent Information Retrieval
 SIAM REVIEW
, 1995
"... Currently, most approaches to retrieving textual materials from scientific databases depend on a lexical match between words in users' requests and those in or assigned to documents in a database. Because of the tremendous diversity in the words people use to describe the same document, lexical ..."
Abstract

Cited by 672 (18 self)
 Add to MetaCart
Currently, most approaches to retrieving textual materials from scientific databases depend on a lexical match between words in users' requests and those in or assigned to documents in a database. Because of the tremendous diversity in the words people use to describe the same document, lexical methods are necessarily incomplete and imprecise. Using the singular value decomposition (SVD), one can take advantage of the implicit higherorder structure in the association of terms with documents by determining the SVD of large sparse term by document matrices. Terms and documents represented by 200300 of the largest singular vectors are then matched against user queries. We call this retrieval method Latent Semantic Indexing (LSI) because the subspace represents important associative relationships between terms and documents that are not evident in individual documents. LSI is a completely automatic yet intelligent indexing method, widely applicable, and a promising way to improve users...
Results 1  10
of
90,527