• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations | Disambiguate

LOCUS: Learning Object Classes with Unsupervised Segmentation (2005)

by J. Winn
Venue:in ICCV
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 195
Next 10 →

TextonBoost: Joint Appearance, Shape and Context Modeling for Multi-Class Object . . .

by J. Shotton, J. Winn, C. Rother, A. Criminisi - IN ECCV , 2006
"... This paper proposes a new approach to learning a discriminative model of object classes, incorporating appearance, shape and context information efficiently. The learned model is used for automatic visual recognition and semantic segmentation of photographs. Our discriminative model exploits nov ..."
Abstract - Cited by 426 (17 self) - Add to MetaCart
This paper proposes a new approach to learning a discriminative model of object classes, incorporating appearance, shape and context information efficiently. The learned model is used for automatic visual recognition and semantic segmentation of photographs. Our discriminative model exploits novel features, based on textons, which jointly model shape and texture. Unary classification and feature selection is achieved using shared boosting to give an efficient classifier which can be applied to a large number of classes. Accurate image segmentation is achieved by incorporating these classifiers in a conditional random field. Efficient training

Learning to detect unseen object classes by betweenclass attribute transfer

by Christoph H. Lampert, Hannes Nickisch, Stefan Harmeling - In CVPR , 2009
"... We study the problem of object classification when training and test classes are disjoint, i.e. no training examples of the target classes are available. This setup has hardly been studied in computer vision research, but it is the rule rather than the exception, because the world contains tens of t ..."
Abstract - Cited by 363 (5 self) - Add to MetaCart
We study the problem of object classification when training and test classes are disjoint, i.e. no training examples of the target classes are available. This setup has hardly been studied in computer vision research, but it is the rule rather than the exception, because the world contains tens of thousands of different object classes and for only a very few of them image, collections have been formed and annotated with suitable class labels. In this paper, we tackle the problem by introducing attribute-based classification. It performs object detection based on a human-specified high-level description of the target objects instead of training images. The description consists of arbitrary semantic attributes, like shape, color or even geographic information. Because such properties transcend the specific learning task at hand, they can be pre-learned, e.g. from image datasets unrelated to the current task. Afterwards, new classes can be detected based on their attribute representation, without the need for a new training phase. In order to evaluate our method and to facilitate research in this area, we have assembled a new largescale dataset, “Animals with Attributes”, of over 30,000 animal images that match the 50 classes in Osherson’s classic table of how strongly humans associate 85 semantic attributes with animal classes. Our experiments show that by using an attribute layer it is indeed possible to build a learning object detection system that does not require any training images of the target classes. 1.
(Show Context)

Citation Context

... representation invariant to non-informative image variations. Similarly, image segmentation has been proposed as an unsupervised method to extract contours that are discriminative for object classes =-=[37]-=-. Such preprocessing steps are generic in the sense that they still allow the subsequent detection of arbitrary object classes. However, the basic elements, local image descriptors or segments shapes,...

TextonBoost for Image Understanding: Multi-Class Object Recognition and Segmentation by Jointly Modeling Texture, Layout, and Context

by Jamie Shotton, John Winn, Carsten Rother, Antonio Criminisi , 2007
"... This paper details a new approach for learning a discriminative model of object classes, incorporating texture, layout, and context information efficiently. The learned model is used for automatic visual understanding and semantic segmentation of photographs. Our discriminative model exploits textur ..."
Abstract - Cited by 217 (9 self) - Add to MetaCart
This paper details a new approach for learning a discriminative model of object classes, incorporating texture, layout, and context information efficiently. The learned model is used for automatic visual understanding and semantic segmentation of photographs. Our discriminative model exploits texture-layout filters, novel features based on textons, which jointly model patterns of texture and their spatial layout. Unary classification and feature selection is achieved using shared boosting to give an efficient classifier which can be applied to a large number of classes. Accurate image segmentation is achieved by incorporating the unary classifier in a conditional random field, which (i) captures the spatial interactions between class labels of neighboring pixels, and (ii) improves the segmentation of specific object instances. Efficient training of the model on large datasets is achieved by exploiting both random feature selection and piecewise training methods. High classification and segmentation accuracy is

Combining top-down and bottom-up segmentation

by Eran Borenstein - In Proceedings IEEE workshop on Perceptual Organization in Computer Vision, CVPR , 2004
"... In this work we show how to combine bottom-up and topdown approaches into a single figure-ground segmentation process. This process provides accurate delineation of object boundaries that cannot be achieved by either the topdown or bottom-up approach alone. The top-down approach uses object represen ..."
Abstract - Cited by 191 (2 self) - Add to MetaCart
In this work we show how to combine bottom-up and topdown approaches into a single figure-ground segmentation process. This process provides accurate delineation of object boundaries that cannot be achieved by either the topdown or bottom-up approach alone. The top-down approach uses object representation learned from examples to detect an object in a given input image and provide an approximation to its figure-ground segmentation. The bottomup approach uses image-based criteria to define coherent groups of pixels that are likely to belong together to either the figure or the background part. The combination provides a final segmentation that draws on the relative merits of both approaches: The result is as close as possible to the top-down approximation, but is also constrained by the bottom-up process to be consistent with significant image discontinuities. We construct a global cost function that represents these top-down and bottom-up requirements. We then show how the global minimum of this function can be efficiently found by applying the sum-product algorithm. This algorithm also provides a confidence map that can be used to identify image regions where additional top-down or bottom-up information may further improve the segmentation. Our experiments show that the results derived from the algorithm are superior to results given by a pure top-down or pure bottom-up approach. The scheme has broad applicability, enabling the combined use of a range of existing bottom-up and top-down segmentations. 1.
(Show Context)

Citation Context

...n called Pictorial Structures (PS) from video sequences. The PS is combined with a contrast dependent Markov Random Field (MRF) that biases the segmentation to follow image boundaries. Winn and Jojic =-=[19]-=- use unsegmented images to learn a global figure/ground mask and a global edge mask that represent the “average” shape and edges of objects in the class. Shape and edge variations are constrained sole...

A.Blake. Cosegmentation of image pairs by histogram matching - incorporating a global constraint into MRFs

by Carsten Rother, Vladimir Kolmogorov, Tom Minka, Andrew Blake - In CVPR , 2006
"... We introduce the term cosegmentation which denotes the task of segmenting simultaneously the common parts of an image pair. A generative model for cosegmentation is presented. Inference in the model leads to minimizing an energy with an MRF term encoding spatial coherency and a global constraint whi ..."
Abstract - Cited by 176 (3 self) - Add to MetaCart
We introduce the term cosegmentation which denotes the task of segmenting simultaneously the common parts of an image pair. A generative model for cosegmentation is presented. Inference in the model leads to minimizing an energy with an MRF term encoding spatial coherency and a global constraint which attempts to match the appearance histograms of the common parts. This energy has not been proposed previously and its optimization is challenging and NP-hard. For this problem a novel optimization scheme which we call trust region graph cuts is presented. We demonstrate that this framework has the potential to improve a wide range of research: Object driven image retrieval, video tracking and segmentation, and interactive image editing. The power of the framework lies in its generality, the common part can be a rigid/non-rigid object (or scene), observed from different viewpoints or even similar objects of the same class. 1.
(Show Context)

Citation Context

... other modalities [13], is normally needed to correct those errors. Another source of information for correcting segmentation is to supply a database of related images and segment them simultaneously =-=[21]-=-. Here we demonstrate that supplying just one additional image can be sufficient to segment both together, to higher accuracy than is achieved with either one alone. Furthermore, in contrast to [21] w...

Decomposing a Scene into Geometric and Semantically Consistent Regions

by Stephen Gould, Richard Fulton, Daphne Koller
"... High-level, or holistic, scene understanding involves reasoning about objects, regions, and the 3D relationships between them. This requires a representation above the level of pixels that can be endowed with high-level attributes such as class of object/region, its orientation, and (rough 3D) locat ..."
Abstract - Cited by 174 (11 self) - Add to MetaCart
High-level, or holistic, scene understanding involves reasoning about objects, regions, and the 3D relationships between them. This requires a representation above the level of pixels that can be endowed with high-level attributes such as class of object/region, its orientation, and (rough 3D) location within the scene. Towards this goal, we propose a region-based model which combines appearance and scene geometry to automatically decompose a scene into semantically meaningful regions. Our model is defined in terms of a unified energy function over scene appearance and structure. We show how this energy function can be learned from data and present an efficient inference technique that makes use of multiple over-segmentations of the image to propose moves in the energy-space. We show, experimentally, that our method achieves state-of-the-art performance on the tasks of both multi-class image segmentation and geometric reasoning. Finally, by understanding region classes and geometry, we show how our model can be used as the basis for 3D reconstruction of the scene. 1.
(Show Context)

Citation Context

...cets of computer vision that have, in recent years, been treated as separate problems. The problem of multi-class image segmentation (or labeling) has been successfully addressed by a number of works =-=[7, 22, 17, 23, 24, 5]-=-. The goal here is to label every pixel in the image with a single class label. Typically these algorithms construct CRFs over the pixels (or small coherent regions called superpixels) with local clas...

V.: What is an object

by Bogdan Alexe, Thomas Deselaers, Vittorio Ferrari , 2010
"... We present a generic objectness measure, quantifying how likely it is for an image window to contain an object of any class. We explicitly train it to distinguish objects with a well-defined boundary in space, such as cows and telephones, from amorphous background elements, such as grass and road. T ..."
Abstract - Cited by 172 (15 self) - Add to MetaCart
We present a generic objectness measure, quantifying how likely it is for an image window to contain an object of any class. We explicitly train it to distinguish objects with a well-defined boundary in space, such as cows and telephones, from amorphous background elements, such as grass and road. The measure combines in a Bayesian framework several image cues measuring characteristics of objects, such as appearing different from their surroundings and having a closed boundary. This includes an innovative cue measuring the closed boundary characteristic. In experiments on the challenging PASCAL VOC 07 dataset, we show this new cue to outperform a state-of-the-art saliency measure [17], and the combined measure to perform better than any cue alone. Finally, we show how to sample windows from an image according to their objectness distribution and give an algorithm to employ them as location priors for modern class-specific object detectors. In experiments on PASCAL VOC 07 we show this greatly reduces the number of windows evaluated by class-specific object detectors. 1.
(Show Context)

Citation Context

... measure can act as a focus of attention mechanism in other applications. It can facilitate learning new classes in a weakly supervised scenario [6], where the location of object instances is unknown =-=[12, 34]-=-. Similarly, it can help object tracking in video, e.g. incorporated as a likelihood term in a CONDENSATION framework [18]. The source code for the objectness measure is available from http://www.visi...

The layout consistent random field for recognizing and segmenting partially occluded objects

by John Winn - In Proceedings of IEEE CVPR , 2006
"... This paper addresses the problem of detecting and segmenting ..."
Abstract - Cited by 152 (8 self) - Add to MetaCart
This paper addresses the problem of detecting and segmenting
(Show Context)

Citation Context

...rts correspond to sparsely detected interest points and so parts are limited in size, cannot represent untextured regions and do not provide a segmentation of the image. More recently, Winn and Jojic =-=[18]-=- used 1 Note that throughout the paper, ‘background’ is used to mean pixels not belonging to an identified object class and ‘foreground’ is used to mean pixels that do belong to the class. Hence it is...

Spatially coherent latent topic model for concurrent object segmentation and classification

by Liangliang Cao, Li Fei-fei - IN: PROCEEDINGS OF IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION , 2007
"... We present a novel generative model for simultaneously recognizing and segmenting object and scene classes. Our model is inspired by the traditional bag of words representation of texts and images as well as a number of related generative models, including probabilistic Latent Sematic Analysis (pL ..."
Abstract - Cited by 150 (3 self) - Add to MetaCart
We present a novel generative model for simultaneously recognizing and segmenting object and scene classes. Our model is inspired by the traditional bag of words representation of texts and images as well as a number of related generative models, including probabilistic Latent Sematic Analysis (pLSA) and Latent Dirichlet Allocation (LDA). A major drawback of the pLSA and LDA models is the assumption that each patch in the image is independently generated given its corresponding latent topic. While such representation provide an efficient computational method, it lacks the power to describe the visually coherent images and scenes. Instead, we propose a spatially coherent latent topic model (Spatial-LTM). Spatial-LTM represents an image containing objects in a hierarchical way by oversegmented image regions of homogeneous appearances and the salient image patches within the regions. Only one single latent topic is assigned to the image patches within each region, enforcing the spatial coherency of the model. This idea gives rise to the following merits of Spatial-LTM: (1) Spatial-LTM provides a unified representation for spatially coherent bag of words topic models; (2) Spatial-LTM can simultaneously segment and classify objects, even in the case of occlusion and multiple instances; and (3) Spatial-LTM can be trained either unsupervised or supervised, as well as when partial object labels are provided. We verify the success of our model in a number of segmentation and classification experiments. E. Coherent regions for
(Show Context)

Citation Context

...TM is tolerant to shape deformation and transformation without sacrificing the computational efficiency to model the spatial freedom. In this way, we differ from the shape based segmentation works in =-=[4, 31, 16]-=- and classification works in [1, 18, 17, 11, 5]. Our work is also related to Russell et. al’s recent work on object segmentation [25]. Both works employ visual words for the task of image segmentation...

An exemplar model for learning object class.

by O Chum, A Zisserman - In CVPR, , 2007
"... ..."
Abstract - Cited by 141 (3 self) - Add to MetaCart
Abstract not found
(Show Context)

Citation Context

...the set of training images in order to “tease out” the class instance in each image. Several previous methods have cast this as an optimization problem, fitting a generative model – for example LOCUS =-=[23]-=- and the Constellation model [8] optimize model likelihood. We also formulate the problem as one of optimization, but with a more general model than those of [8, 23]. LOCUS is limited by its use of th...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University