Results 11  20
of
766
Rational approximations to rational models: Alternative algorithms for category learning
"... Rational models of cognition typically consider the abstract computational problems posed by the environment, assuming that people are capable of optimally solving those problems. This differs from more traditional formal models of cognition, which focus on the psychological processes responsible fo ..."
Abstract

Cited by 58 (16 self)
 Add to MetaCart
(Show Context)
Rational models of cognition typically consider the abstract computational problems posed by the environment, assuming that people are capable of optimally solving those problems. This differs from more traditional formal models of cognition, which focus on the psychological processes responsible for behavior. A basic challenge for rational models is thus explaining how optimal solutions can be approximated by psychological processes. We outline a general strategy for answering this question, namely to explore the psychological plausibility of approximation algorithms developed in computer science and statistics. In particular, we argue that Monte Carlo methods provide a source of “rational process models” that connect optimal solutions to psychological processes. We support this argument through a detailed example, applying this approach to Anderson’s (1990, 1991) Rational Model of Categorization (RMC), which involves a particularly challenging computational problem. Drawing on a connection between the RMC and ideas from nonparametric Bayesian statistics, we propose two alternative algorithms for approximate inference in this model. The algorithms we consider include Gibbs sampling, a procedure
Robust MultiPerson Tracking from a Mobile Platform
"... In this paper, we address the problem of multiperson tracking in busy pedestrian zones using a stereo rig mounted on a mobile platform. The complexity of the problem calls for an integrated solution that extracts as much visual information as possible and combines it through cognitive feedback cycl ..."
Abstract

Cited by 55 (12 self)
 Add to MetaCart
In this paper, we address the problem of multiperson tracking in busy pedestrian zones using a stereo rig mounted on a mobile platform. The complexity of the problem calls for an integrated solution that extracts as much visual information as possible and combines it through cognitive feedback cycles. We propose such an approach, which jointly estimates camera position, stereo depth, object detection, and tracking. The interplay between those components is represented by a graphical model. Since the model has to incorporate objectobject interactions and temporal links to past frames, direct inference is intractable. We therefore propose a twostage procedure: for each frame we first solve a simplified version of the model (disregarding interactions and temporal continuity) to estimate the scene geometry and an overcomplete set of object detections. Conditioned on these results, we then address object interactions, tracking, and prediction in a second step. The approach is experimentally evaluated on several long and difficult video sequences from busy innercity locations. Our results show that the proposed integration makes it possible to deliver robust tracking performance in scenes of realistic complexity.
Bayesian models of cognition
"... For over 200 years, philosophers and mathematicians have been using probability theory to describe human cognition. While the theory of probabilities was first developed as a means of analyzing games of chance, it quickly took on a larger and deeper significance as a formal account of how rational a ..."
Abstract

Cited by 48 (3 self)
 Add to MetaCart
(Show Context)
For over 200 years, philosophers and mathematicians have been using probability theory to describe human cognition. While the theory of probabilities was first developed as a means of analyzing games of chance, it quickly took on a larger and deeper significance as a formal account of how rational agents should reason in situations of uncertainty
Efficient Highly OverComplete Sparse Coding using a Mixture Model
"... Abstract. Sparse coding of sensory data has recently attracted notable attention in research of learning useful features from the unlabeled data. Empirical studies show that mapping the data into a significantly higherdimensional space with sparse coding can lead to superior classification performan ..."
Abstract

Cited by 45 (0 self)
 Add to MetaCart
(Show Context)
Abstract. Sparse coding of sensory data has recently attracted notable attention in research of learning useful features from the unlabeled data. Empirical studies show that mapping the data into a significantly higherdimensional space with sparse coding can lead to superior classification performance. However, computationally it is challenging to learn a set of highly overcomplete dictionary bases and to encode the test data with the learned bases. In this paper, we describe a mixture sparse coding model that can produce highdimensional sparse representations very efficiently. Besides the computational advantage, the model effectively encourages data that are similar to each other to enjoy similar sparse representations. What’s more, the proposed model can be regarded as an approximation to the recently proposed local coordinate coding (LCC), which states that sparse coding can approximately learn the nonlinear manifold of the sensory data in a locally linear manner. Therefore, the feature learned by the mixture sparse coding model works pretty well with linear classifiers. We apply the proposed model to PASCAL VOC 2007 and 2009 datasets for the classification task, both achieving stateoftheart performances. Key words: Sparse coding, highly overcomplete dictionary training, mixture model, mixture sparse coding, image classification, PASCAL VOC challenge 1
Geos: Geodesic image segmentation
 ECCV '08 PROCEEDINGS OF THE 10TH EUROPEAN CONFERENCE ON COMPUTER VISION: PART I
, 2008
"... Abstract. This paper presents GeoS, a new algorithm for the efficient segmentation of ndimensional image and video data. The segmentation problem is cast as approximate energy minimization in a conditional random field. A new, parallel filtering operator built upon efficient geodesic distance compu ..."
Abstract

Cited by 43 (4 self)
 Add to MetaCart
(Show Context)
Abstract. This paper presents GeoS, a new algorithm for the efficient segmentation of ndimensional image and video data. The segmentation problem is cast as approximate energy minimization in a conditional random field. A new, parallel filtering operator built upon efficient geodesic distance computation is used to propose a set of spatially smooth, contrastsensitive segmentation hypotheses. An economical search algorithm finds the solution with minimum energy within a sensible and highly restricted subset of all possible labellings. Advantages include: i) computational efficiency with high segmentation accuracy; ii) the ability to estimate an approximation to the posterior over segmentations; iii) the ability to handle generally complex energy models. Comparison with maxflow indicates up to 60 times greater computational efficiency as well as greater memory efficiency. GeoS is validated quantitatively and qualitatively by thorough comparative experiments on existing and novel groundtruth data. Numerous results on interactive and automatic segmentation of photographs, video and volumetric medical image data are presented. 1
Did You See Bob?: Human Localization using Mobile Phones
"... Finding a person in a public place, such as in a library, conference hotel, or shopping mall, can be difficult. The difficulty arises from not knowing where the person may be at that time; even if known, navigating through an unfamiliar place may be frustrating. Maps and floor plans help in some occ ..."
Abstract

Cited by 41 (0 self)
 Add to MetaCart
(Show Context)
Finding a person in a public place, such as in a library, conference hotel, or shopping mall, can be difficult. The difficulty arises from not knowing where the person may be at that time; even if known, navigating through an unfamiliar place may be frustrating. Maps and floor plans help in some occasions, but such maps may not be always handy. In a small scale poll, 80 % of users responded that the ideal solution would be “to have an escort walk me to the desired person”. This paper identifies the possibility of using mobile phone sensors and opportunistic userintersections to develop an electronic escort service. By periodically learning the walking trails of different individuals, as well as how they encounter each other in spacetime, a route can be computed between any pair of persons. The problem bears resemblance to routing packets in delay tolerant networks, however, its application in the context of human localization raises distinct research challenges. We design and implement Escort, a system that guides a user to the vicinity of a desired person in a public place. We only use an audio beacon, randomly placed in the building, to enable a reference frame. We do not rely on GPS, WiFi, or wardriving to locate a person – the Escort user only needs to follow an arrow displayed on the phone. Evaluation results from experiments in parking lots and university buildings show that, on average, the user is brought to within 8m of the destination. We believe this is an encouraging result, opening new possibilities in mobile, social localization.
Automatic classification of MR scans in Alzheimer's disease
 Brain
, 2008
"... These authors contributed equally to this work. To be diagnostically useful, structural MRI must reliably distinguish Alzheimer’s disease (AD) from normal aging in individual scans. Recent advances in statistical learning theory have led to the application of support vector machines to MRI for detec ..."
Abstract

Cited by 40 (0 self)
 Add to MetaCart
(Show Context)
These authors contributed equally to this work. To be diagnostically useful, structural MRI must reliably distinguish Alzheimer’s disease (AD) from normal aging in individual scans. Recent advances in statistical learning theory have led to the application of support vector machines to MRI for detection of a variety of disease states.The aims of this study were to assess how successfully support vector machines assigned individual diagnoses and to determine whether datasets combined from multiple scanners and different centres could be used to obtain effective classification of scans. We used linear support vector machines to classify the grey matter segment of T1weighted MR scans from pathologically proven AD patients and cognitively normal elderly individuals obtained from two centres with different scanning equipment. Because the clinical diagnosis of mild AD is difficult we also tested the ability of support vector machines to differentiate control scans from patients without postmortem confirmation. Finally we sought to use these methods to differentiate scans between patients suffering from AD from those with frontotemporal lobar degeneration.Up to 96 % of pathologically verified AD patients were correctly classified using whole brain
Statistical analysis on stiefel and grassmann manifolds with applications in computer vision
 CVPR
, 2008
"... Many applications in computer vision and pattern recognition involve drawing inferences on certain manifoldvalued parameters. In order to develop accurate inference algorithms on these manifolds we need to a) understand the geometric structure of these manifolds b) derive appropriate distance measu ..."
Abstract

Cited by 39 (6 self)
 Add to MetaCart
(Show Context)
Many applications in computer vision and pattern recognition involve drawing inferences on certain manifoldvalued parameters. In order to develop accurate inference algorithms on these manifolds we need to a) understand the geometric structure of these manifolds b) derive appropriate distance measures and c) develop probability distribution functions (pdf) and estimation techniques that are consistent with the geometric structure of these manifolds. In this paper, we consider two related manifolds the Stiefel manifold and the Grassmann manifold, which arise naturally in several vision applications such as spatiotemporal modeling, affine invariant shape analysis, image matching and learning theory. We show how accurate statistical characterization that reflects the geometry of these manifolds allows us to design efficient algorithms that compare favorably to the state of the art in these very different applications. In particular, we describe appropriate distance measures and parametric and nonparametric density estimators on these manifolds. These methods are then used to learn class conditional densities for applications such as activity recognition, video based face recognition and shape classification.
Tied factor analysis for face recognition across large pose changes
 BMVC
, 2006
"... Abstract—Face recognition algorithms perform very unreliably when the pose of the probe face is different from the gallery face: typical feature vectors vary more with pose than with identity. We propose a generative model that creates a onetomany mapping from an idealized “identity ” space to the ..."
Abstract

Cited by 37 (4 self)
 Add to MetaCart
Abstract—Face recognition algorithms perform very unreliably when the pose of the probe face is different from the gallery face: typical feature vectors vary more with pose than with identity. We propose a generative model that creates a onetomany mapping from an idealized “identity ” space to the observed data space. In identity space, the representation for each individual does not vary with pose. We model the measured feature vector as being generated by a posecontingent linear transformation of the identity variable in the presence of Gaussian noise. We term this model “tied ” factor analysis. The choice of linear transformation (factors) depends on the pose, but the loadings are constant (tied) for a given individual. We use the EM algorithm to estimate the linear transformations and the noise parameters from training data. We propose a probabilistic distance metric that allows a full posterior over possible matches to be established. We introduce a novel feature extraction process and investigate recognition performance by using the FERET, XM2VTS, and PIE databases. Recognition performance compares favorably with contemporary approaches. Index Terms—Computing methodologies, pattern recognition, applications, face and gesture recognition. Ç 1
Structured learning and prediction in computer vision
 IN FOUNDATIONS AND TRENDS IN COMPUTER GRAPHICS AND VISION
, 2010
"... ..."
(Show Context)