Results 11  20
of
1,841,068
A theory of memory retrieval
 PSYCHOL. REV
, 1978
"... A theory of memory retrieval is developed and is shown to apply over a range of experimental paradigms. Access to memory traces is viewed in terms of a resonance metaphor. The probe item evokes the search set on the basis of probememory item relatedness, just as a ringing tuning fork evokes sympath ..."
Abstract

Cited by 731 (82 self)
 Add to MetaCart
sympathetic vibrations in other tuning forks. Evidence is accumulated in parallel from each probememory item comparison, and each comparison is modeled by a continuous random walk process. In item recognition, the decision process is selfterminating on matching comparisons and exhaustive on nonmatching
Search and replication in unstructured peertopeer networks
, 2002
"... Abstract Decentralized and unstructured peertopeer networks such as Gnutella are attractive for certain applicationsbecause they require no centralized directories and no precise control over network topologies and data placement. However, the floodingbased query algorithm used in Gnutella does n ..."
Abstract

Cited by 671 (6 self)
 Add to MetaCart
propose a query algorithm based on multiple random walks that resolves queries almost as quickly as gnutella's flooding method while reducing the network traffic by two orders of magnitude in many cases. We also present a distributed replication strategy that yields closetooptimal performance
Image denoising using a scale mixture of Gaussians in the wavelet domain
 IEEE TRANS IMAGE PROCESSING
, 2003
"... We describe a method for removing noise from digital images, based on a statistical model of the coefficients of an overcomplete multiscale oriented basis. Neighborhoods of coefficients at adjacent positions and scales are modeled as the product of two independent random variables: a Gaussian vecto ..."
Abstract

Cited by 509 (17 self)
 Add to MetaCart
We describe a method for removing noise from digital images, based on a statistical model of the coefficients of an overcomplete multiscale oriented basis. Neighborhoods of coefficients at adjacent positions and scales are modeled as the product of two independent random variables: a Gaussian
A New Method for Solving Hard Satisfiability Problems
 AAAI
, 1992
"... We introduce a greedy local search procedure called GSAT for solving propositional satisfiability problems. Our experiments show that this procedure can be used to solve hard, randomly generated problems that are an order of magnitude larger than those that can be handled by more traditional approac ..."
Abstract

Cited by 725 (21 self)
 Add to MetaCart
discussed. GSAT is best viewed as a modelfinding procedure. Its good performance suggests that it may be advantageous to reformulate reasoning tasks that have traditionally been viewed as theoremproving problems as modelfinding tasks.
The unity and diversity of executive functions and their contributions to complex “Frontal Lobe” tasks: a latent variable analysis
 COGNIT PSYCHOL
, 2000
"... This individual differences study examined the separability of three often postulated executive functions—mental set shifting ("Shifting"), information updating and monitoring ("Updating"), and inhibition of prepotent responses ("Inhibition")—and their roles in complex ..."
Abstract

Cited by 631 (9 self)
 Add to MetaCart
Sorting Test (WCST), Tower of Hanoi (TOH), random number generation (RNG), operation span, and dual tasking. Confirmatory factor analysis indicated that the three target executive functions are moderately correlated with one another, but are clearly separable. Moreover, structural equation modeling
The strength of weak learnability
 MACHINE LEARNING
, 1990
"... This paper addresses the problem of improving the accuracy of an hypothesis output by a learning algorithm in the distributionfree (PAC) learning model. A concept class is learnable (or strongly learnable) if, given access to a Source of examples of the unknown concept, the learner with high prob ..."
Abstract

Cited by 852 (24 self)
 Add to MetaCart
probability is able to output an hypothesis that is correct on all but an arbitrarily small fraction of the instances. The concept class is weakly learnable if the learner can produce an hypothesis that performs only slightly better than random guessing. In this paper, it is shown that these two notions
Dynamic Bayesian Networks: Representation, Inference and Learning
, 2002
"... Modelling sequential data is important in many areas of science and engineering. Hidden Markov models (HMMs) and Kalman filter models (KFMs) are popular for this because they are simple and flexible. For example, HMMs have been used for speech recognition and biosequence analysis, and KFMs have bee ..."
Abstract

Cited by 755 (3 self)
 Add to MetaCart
random variable. DBNs generalize KFMs by allowing arbitrary probability distributions, not just (unimodal) linearGaussian. In this thesis, I will discuss how to represent many different kinds of models as DBNs, how to perform exact and approximate inference in DBNs, and how to learn DBN models from
Compressed sensing
, 2004
"... We study the notion of Compressed Sensing (CS) as put forward in [14] and related work [20, 3, 4]. The basic idea behind CS is that a signal or image, unknown but supposed to be compressible by a known transform, (eg. wavelet or Fourier), can be subjected to fewer measurements than the nominal numbe ..."
Abstract

Cited by 3559 (22 self)
 Add to MetaCart
number of pixels, and yet be accurately reconstructed. The samples are nonadaptive and measure ‘random’ linear combinations of the transform coefficients. Approximate reconstruction is obtained by solving for the transform coefficients consistent with measured data and having the smallest possible `1
Robust face recognition via sparse representation
 IEEE TRANS. PATTERN ANALYSIS AND MACHINE INTELLIGENCE
, 2008
"... We consider the problem of automatically recognizing human faces from frontal views with varying expression and illumination, as well as occlusion and disguise. We cast the recognition problem as one of classifying among multiple linear regression models, and argue that new theory from sparse signa ..."
Abstract

Cited by 912 (40 self)
 Add to MetaCart
is correctly computed. Unconventional features such as downsampled images and random projections perform just as well as conventional features such as Eigenfaces and Laplacianfaces, as long as the dimension of the feature space surpasses certain threshold, predicted by the theory of sparse representation
On the time course of perceptual choice: the leaky competing accumulator model
 PSYCHOLOGICAL REVIEW
, 2001
"... The time course of perceptual choice is discussed in a model based on gradual and stochastic accumulation of information in nonlinear decision units with leakage (or decay of activation) and competition through lateral inhibition. In special cases, the model becomes equivalent to a classical diffus ..."
Abstract

Cited by 460 (19 self)
 Add to MetaCart
diffusion process, but leakage and mutual inhibition work together to address several challenges to existing diffusion, randomwalk, and accumulator models. The model provides a good account of data from choice tasks using both timecontrolled (e.g., deadline or response signal) and standard reaction time
Results 11  20
of
1,841,068