Results 1  10
of
49,733
A Combinatorial Formula for Principal Minors of a Matrix with Treemetric Exponents and Its Applications
, 2013
"... ar ..."
A Guided Tour to Approximate String Matching
 ACM COMPUTING SURVEYS
, 1999
"... We survey the current techniques to cope with the problem of string matching allowing errors. This is becoming a more and more relevant issue for many fast growing areas such as information retrieval and computational biology. We focus on online searching and mostly on edit distance, explaining t ..."
Abstract

Cited by 584 (38 self)
 Add to MetaCart
the problem and its relevance, its statistical behavior, its history and current developments, and the central ideas of the algorithms and their complexities. We present a number of experiments to compare the performance of the different algorithms and show which are the best choices according to each case
Survey on Independent Component Analysis
 NEURAL COMPUTING SURVEYS
, 1999
"... A common problem encountered in such disciplines as statistics, data analysis, signal processing, and neural network research, is nding a suitable representation of multivariate data. For computational and conceptual simplicity, such a representation is often sought as a linear transformation of the ..."
Abstract

Cited by 2241 (104 self)
 Add to MetaCart
of the original data. Wellknown linear transformation methods include, for example, principal component analysis, factor analysis, and projection pursuit. A recently developed linear transformation method is independent component analysis (ICA), in which the desired representation is the one that minimizes
FAST VOLUME RENDERING USING A SHEARWARP FACTORIZATION OF THE VIEWING TRANSFORMATION
, 1995
"... Volume rendering is a technique for visualizing 3D arrays of sampled data. It has applications in areas such as medical imaging and scientific visualization, but its use has been limited by its high computational expense. Early implementations of volume rendering used bruteforce techniques that req ..."
Abstract

Cited by 541 (2 self)
 Add to MetaCart
Volume rendering is a technique for visualizing 3D arrays of sampled data. It has applications in areas such as medical imaging and scientific visualization, but its use has been limited by its high computational expense. Early implementations of volume rendering used bruteforce techniques
Graphical models, exponential families, and variational inference
, 2008
"... The formalism of probabilistic graphical models provides a unifying framework for capturing complex dependencies among random variables, and building largescale multivariate statistical models. Graphical models have become a focus of research in many statistical, computational and mathematical fiel ..."
Abstract

Cited by 800 (26 self)
 Add to MetaCart
The formalism of probabilistic graphical models provides a unifying framework for capturing complex dependencies among random variables, and building largescale multivariate statistical models. Graphical models have become a focus of research in many statistical, computational and mathematical
A Signal Processing Approach To Fair Surface Design
, 1995
"... In this paper we describe a new tool for interactive freeform fair surface design. By generalizing classical discrete Fourier analysis to twodimensional discrete surface signals  functions defined on polyhedral surfaces of arbitrary topology , we reduce the problem of surface smoothing, or fai ..."
Abstract

Cited by 668 (15 self)
 Add to MetaCart
In this paper we describe a new tool for interactive freeform fair surface design. By generalizing classical discrete Fourier analysis to twodimensional discrete surface signals  functions defined on polyhedral surfaces of arbitrary topology , we reduce the problem of surface smoothing
Panel Cointegration; Asymptotic and Finite Sample Properties of Pooled Time Series Tests, With an Application to the PPP Hypothesis; New Results. Working paper
, 1997
"... We examine properties of residualbased tests for the null of no cointegration for dynamic panels in which both the shortrun dynamics and the longrun slope coefficients are permitted to be heterogeneous across individual members of the panel+ The tests also allow for individual heterogeneous fixed ..."
Abstract

Cited by 499 (13 self)
 Add to MetaCart
We examine properties of residualbased tests for the null of no cointegration for dynamic panels in which both the shortrun dynamics and the longrun slope coefficients are permitted to be heterogeneous across individual members of the panel+ The tests also allow for individual heterogeneous fixed effects and trend terms, and we consider both pooled within dimension tests and group mean between dimension tests+ We derive limiting distributions for these and show that they are normal and free of nuisance parameters+ We also provide Monte Carlo evidence to demonstrate their small sample size and power performance, and we illustrate their use in testing purchasing power parity for the post–Bretton Woods period+ 1.
String theory and noncommutative geometry
 JHEP
, 1999
"... We extend earlier ideas about the appearance of noncommutative geometry in string theory with a nonzero Bfield. We identify a limit in which the entire string dynamics is described by a minimally coupled (supersymmetric) gauge theory on a noncommutative space, and discuss the corrections away from ..."
Abstract

Cited by 801 (8 self)
 Add to MetaCart
this limit. Our analysis leads us to an equivalence between ordinary gauge fields and noncommutative gauge fields, which is realized by a change of variables that can be described explicitly. This change of variables is checked by comparing the ordinary DiracBornInfeld theory with its noncommutative
Bayes Factors
, 1995
"... In a 1935 paper, and in his book Theory of Probability, Jeffreys developed a methodology for quantifying the evidence in favor of a scientific theory. The centerpiece was a number, now called the Bayes factor, which is the posterior odds of the null hypothesis when the prior probability on the null ..."
Abstract

Cited by 1766 (74 self)
 Add to MetaCart
In a 1935 paper, and in his book Theory of Probability, Jeffreys developed a methodology for quantifying the evidence in favor of a scientific theory. The centerpiece was a number, now called the Bayes factor, which is the posterior odds of the null hypothesis when the prior probability on the null
Near Optimal Signal Recovery From Random Projections: Universal Encoding Strategies?
, 2004
"... Suppose we are given a vector f in RN. How many linear measurements do we need to make about f to be able to recover f to within precision ɛ in the Euclidean (ℓ2) metric? Or more exactly, suppose we are interested in a class F of such objects— discrete digital signals, images, etc; how many linear m ..."
Abstract

Cited by 1513 (20 self)
 Add to MetaCart
Suppose we are given a vector f in RN. How many linear measurements do we need to make about f to be able to recover f to within precision ɛ in the Euclidean (ℓ2) metric? Or more exactly, suppose we are interested in a class F of such objects— discrete digital signals, images, etc; how many linear
Results 1  10
of
49,733