Results 1  10
of
180,887
Bayesian Interpolation
 Neural Computation
, 1991
"... Although Bayesian analysis has been in use since Laplace, the Bayesian method of modelcomparison has only recently been developed in depth. In this paper, the Bayesian approach to regularisation and modelcomparison is demonstrated by studying the inference problem of interpolating noisy data. T ..."
Abstract

Cited by 721 (17 self)
 Add to MetaCart
Although Bayesian analysis has been in use since Laplace, the Bayesian method of modelcomparison has only recently been developed in depth. In this paper, the Bayesian approach to regularisation and modelcomparison is demonstrated by studying the inference problem of interpolating noisy data
View Interpolation for Image Synthesis
"... Imagespace simplifications have been used to accelerate the calculation of computer graphic images since the dawn of visual simulation. Texture mapping has been used to provide a means by which images may themselves be used as display primitives. The work reported by this paper endeavors to carry t ..."
Abstract

Cited by 605 (0 self)
 Add to MetaCart
Imagespace simplifications have been used to accelerate the calculation of computer graphic images since the dawn of visual simulation. Texture mapping has been used to provide a means by which images may themselves be used as display primitives. The work reported by this paper endeavors to carry
Qualitative Simulation
 Artificial Intelligence
, 2001
"... Qualitative simulation predicts the set of possible behaviors... ..."
Abstract

Cited by 513 (31 self)
 Add to MetaCart
Qualitative simulation predicts the set of possible behaviors...
Ptolemy: A Framework for Simulating and Prototyping Heterogeneous Systems
, 1992
"... Ptolemy is an environment for simulation and prototyping of heterogeneous systems. It uses modern objectoriented software technology (C++) to model each subsystem in a natural and efficient manner, and to integrate these subsystems into a whole. Ptolemy encompasses practically all aspects of design ..."
Abstract

Cited by 569 (90 self)
 Add to MetaCart
of designing signal processing and communications systems, ranging from algorithms and communication strategies, simulation, hardware and software design, parallel computing, and generating realtime prototypes. To accommodate this breadth, Ptolemy must support a plethora of widelydiffering design styles
The Lifting Scheme: A Construction Of Second Generation Wavelets
, 1997
"... . We present the lifting scheme, a simple construction of second generation wavelets, wavelets that are not necessarily translates and dilates of one fixed function. Such wavelets can be adapted to intervals, domains, surfaces, weights, and irregular samples. We show how the lifting scheme leads to ..."
Abstract

Cited by 541 (16 self)
 Add to MetaCart
. We present the lifting scheme, a simple construction of second generation wavelets, wavelets that are not necessarily translates and dilates of one fixed function. Such wavelets can be adapted to intervals, domains, surfaces, weights, and irregular samples. We show how the lifting scheme leads
Halfa century of research on the Stroop effect: An integrative review
 PsychologicalBulletin
, 1991
"... The literature on interference in the Stroop ColorWord Task, covering over 50 years and some 400 studies, is organized and reviewed. In so doing, a set ofl 8 reliable empirical findings is isolated that must be captured by any successful theory of the Stroop effect. Existing theoretical positions a ..."
Abstract

Cited by 621 (14 self)
 Add to MetaCart
The literature on interference in the Stroop ColorWord Task, covering over 50 years and some 400 studies, is organized and reviewed. In so doing, a set ofl 8 reliable empirical findings is isolated that must be captured by any successful theory of the Stroop effect. Existing theoretical positions are summarized and evaluated in view of this critical evidence and the 2 major candidate theoriesrelative speed of processing and automaticity of readingare found to be wanting. It is concluded that recent theories placing the explanatory weight on parallel processing of the irrelevant and the relevant dimensions are likely to be more successful than are earlier theories attempting to locate a single bottleneck in attention. In 1935, J. R. Stroop published his landmark article on attention and interference, an article more influential now than it was then. Why has the Stroop task continued to fascinate us? Perhaps the task is seen as tapping into the primitive operations of cognition, offering clues to the fundamental process of attention. Perhaps the robustness of the phenomenon provides a special challenge to decipher. Together these are powerful attractions
Static Scheduling of Synchronous Data Flow Programs for Digital Signal Processing
 IEEE TRANSACTIONS ON COMPUTERS
, 1987
"... Large grain data flow (LGDF) programming is natural and convenient for describing digital signal processing (DSP) systems, but its runtime overhead is costly in real time or costsensitive applications. In some situations, designers are not willing to squander computing resources for the sake of pro ..."
Abstract

Cited by 592 (37 self)
 Add to MetaCart
Large grain data flow (LGDF) programming is natural and convenient for describing digital signal processing (DSP) systems, but its runtime overhead is costly in real time or costsensitive applications. In some situations, designers are not willing to squander computing resources for the sake
Near Optimal Signal Recovery From Random Projections: Universal Encoding Strategies?
, 2004
"... Suppose we are given a vector f in RN. How many linear measurements do we need to make about f to be able to recover f to within precision ɛ in the Euclidean (ℓ2) metric? Or more exactly, suppose we are interested in a class F of such objects— discrete digital signals, images, etc; how many linear m ..."
Abstract

Cited by 1513 (20 self)
 Add to MetaCart
Suppose we are given a vector f in RN. How many linear measurements do we need to make about f to be able to recover f to within precision ɛ in the Euclidean (ℓ2) metric? Or more exactly, suppose we are interested in a class F of such objects— discrete digital signals, images, etc; how many linear
LanguageBased InformationFlow Security
 IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS
, 2003
"... Current standard security practices do not provide substantial assurance that the endtoend behavior of a computing system satisfies important security policies such as confidentiality. An endtoend confidentiality policy might assert that secret input data cannot be inferred by an attacker throug ..."
Abstract

Cited by 821 (57 self)
 Add to MetaCart
Current standard security practices do not provide substantial assurance that the endtoend behavior of a computing system satisfies important security policies such as confidentiality. An endtoend confidentiality policy might assert that secret input data cannot be inferred by an attacker through the attacker's observations of system output; this policy regulates information flow.
Just Relax: Convex Programming Methods for Identifying Sparse Signals in Noise
, 2006
"... This paper studies a difficult and fundamental problem that arises throughout electrical engineering, applied mathematics, and statistics. Suppose that one forms a short linear combination of elementary signals drawn from a large, fixed collection. Given an observation of the linear combination that ..."
Abstract

Cited by 496 (2 self)
 Add to MetaCart
This paper studies a difficult and fundamental problem that arises throughout electrical engineering, applied mathematics, and statistics. Suppose that one forms a short linear combination of elementary signals drawn from a large, fixed collection. Given an observation of the linear combination
Results 1  10
of
180,887