• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Modelling, visualising and summarising documents with a single convolutional neural network. arXiv preprint arXiv:1406.3830 (2014)

by M Denil, A Demiraj, N Kalchbrenner, P Blunsom, N de Freitas
Add To MetaCart

Tools

Sorted by:
Results 1 - 5 of 5

From Captions to Visual Concepts and Back

by Hao Fang, Li Deng, Jianfeng Gao, Xiaodong He, Margaret Mitchell, John C. Platt, C. Lawrence Zitnick , Geoffrey Zweig, et al. , 2014
"... This paper presents a novel approach for automatically generating image descriptions: visual detectors and language models learn directly from a dataset of image captions. We use Multiple Instance Learning to train visual detectors for words that commonly occur in captions, including many different ..."
Abstract - Cited by 15 (1 self) - Add to MetaCart
This paper presents a novel approach for automatically generating image descriptions: visual detectors and language models learn directly from a dataset of image captions. We use Multiple Instance Learning to train visual detectors for words that commonly occur in captions, including many different parts of speech such as nouns, verbs, and adjectives. The word detector outputs serve as conditional inputs to a maximum-entropy language model. The language model learns from a set of over 400,000 image descriptions to capture the statistics of word usage. We capture global semantics by re-ranking caption candidates using sentence-level features and a deep multimodal similarity model. When human judges compare the system captions to ones written by other people, the system captions have equal or better quality over 23 % of the time.

Learning to rank short text pairs with convolutional deep neural networks

by Aliaksei Severyn, Google Inc, Alessandro Moschitti - In SIGIR , 2015
"... Learning a similarity function between pairs of objects is at the core of learning to rank approaches. In information retrieval tasks we typically deal with query-document pairs, in question answering – question-answer pairs. However, before learning can take place, such pairs needs to be mapped fro ..."
Abstract - Cited by 2 (1 self) - Add to MetaCart
Learning a similarity function between pairs of objects is at the core of learning to rank approaches. In information retrieval tasks we typically deal with query-document pairs, in question answering – question-answer pairs. However, before learning can take place, such pairs needs to be mapped from the original space of symbolic words into some feature space encoding various aspects of their relatedness, e.g. lexical, syntactic and semantic. Feature engineer-ing is often a laborious task and may require external knowledge sources that are not always available or difficult to obtain. Recently, deep learning approaches have gained a lot of attention from the research community and industry for their ability to automatically learn optimal feature representation for a given task, while claim-ing state-of-the-art performance in many tasks in computer vision, speech recognition and natural language processing. In this paper,
(Show Context)

Citation Context

...ng the word embeddings for a given task should be at least in the order of hundred thousands, while in our case the number of query-document pairs is one order of magnitude smaller. Hence, similar to =-=[11, 19, 38]-=- we keep the word embeddings fixed and initialize the word matrix W from an unsupervised neural language model. We choose the dimensionality of our word embeddings to be 50 to be on the line with the ...

Diversifying Restricted Boltzmann Machine for Document Modeling

by Pengtao Xie, Yuntian Deng, Eric P. Xing
"... Restricted Boltzmann Machine (RBM) has shown great ef-fectiveness in document modeling. It utilizes hidden units to discover the latent topics and can learn compact semantic representations for documents which greatly facilitate doc-ument retrieval, clustering and classification. The popular-ity (or ..."
Abstract - Cited by 1 (1 self) - Add to MetaCart
Restricted Boltzmann Machine (RBM) has shown great ef-fectiveness in document modeling. It utilizes hidden units to discover the latent topics and can learn compact semantic representations for documents which greatly facilitate doc-ument retrieval, clustering and classification. The popular-ity (or frequency) of topics in text corpora usually follow a power-law distribution where a few dominant topics occur very frequently while most topics (in the long-tail region) have low probabilities. Due to this imbalance, RBM tends to learn multiple redundant hidden units to best represent dominant topics and ignore those in the long-tail region, which renders the learned representations to be redundant and non-informative. To solve this problem, we propose Di-versified RBM (DRBM) which diversifies the hidden units, to make them cover not only the dominant topics, but also those in the long-tail region. We define a diversity metric and use it as a regularizer to encourage the hidden units to be diverse. Since the diversity metric is hard to optimize directly, we instead optimize its lower bound and prove that maximizing the lower bound with projected gradient ascent can increase this diversity metric. Experiments on docu-ment retrieval and clustering demonstrate that with diver-sification, the document modeling power of DRBM can be greatly improved.

Generating Sentences from Semantic Vector Space Representations

by Mohit Iyyer, Jordan Boyd-graber, Hal Daume ́ Iii
"... Distributed vector space models have recently shown success at capturing the semantic meanings of words [2, 15, 14], phrases and sentences [18, 16, 12], and even full documents [13, 3]. However, there has not been much work in the reverse direction: given a single vector that represents some meaning ..."
Abstract - Add to MetaCart
Distributed vector space models have recently shown success at capturing the semantic meanings of words [2, 15, 14], phrases and sentences [18, 16, 12], and even full documents [13, 3]. However, there has not been much work in the reverse direction: given a single vector that represents some meaning, can we generate grammatically correct text that retains that
(Show Context)

Citation Context

...ado.edu 1 Introduction Distributed vector space models have recently shown success at capturing the semantic meanings of words [2, 15, 14], phrases and sentences [18, 16, 12], and even full documents =-=[13, 3]-=-. However, there has not been much work in the reverse direction: given a single vector that represents some meaning, can we generate grammatically correct text that retains that meaning? The first wo...

Deep Multi-Instance Transfer Learning

by Dimitrios Kotzias, Misha Denil Phil Blunsom, O De Freitas
"... We present a new approach for transferring knowledge from groups to individu-als that comprise them. We evaluate our method in text, by inferring the ratings of individual sentences using full-review ratings. This approach combines ideas from transfer learning, deep learning and multi-instance learn ..."
Abstract - Add to MetaCart
We present a new approach for transferring knowledge from groups to individu-als that comprise them. We evaluate our method in text, by inferring the ratings of individual sentences using full-review ratings. This approach combines ideas from transfer learning, deep learning and multi-instance learning, and reduces the need for laborious human labelling of fine-grained data when abundant labels are available at the group level. 1
(Show Context)

Citation Context

...stance transfer learning. The first step in this approach involves creating a representation for sentences. We do that by training the supervised document convolutional neural network of Denil et al. =-=[9]-=- to predict review scores. As a result of this training, we obtain embeddings (vectors in a metric space) for words, sentences and reviews. These embeddings are the features for the individuals (sente...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University