• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Simlex-999: Evaluating semantic models with (genuine) similarity estimation. arXiv preprint arXiv:1408.3456. (2014)

by F Hill, R Reichart, A Korhonen
Add To MetaCart

Tools

Sorted by:
Results 1 - 9 of 9

Reserating the awesometastic: An automatic extension of the WordNet taxonomy for novel terms

by David Jurgens, Mohammad Taher Pilehvar
"... This paper presents CROWN, an automatically con-structed extension of WordNet that augments its taxonomy with novel lemmas from Wiktionary. CROWN fills the important gap in WordNet’s lexi-con for slang, technical, and rare lemmas, and more than doubles its current size. In two evaluations, we demons ..."
Abstract - Cited by 1 (1 self) - Add to MetaCart
This paper presents CROWN, an automatically con-structed extension of WordNet that augments its taxonomy with novel lemmas from Wiktionary. CROWN fills the important gap in WordNet’s lexi-con for slang, technical, and rare lemmas, and more than doubles its current size. In two evaluations, we demonstrate that the construction procedure is accu-rate and has a significant impact on a WordNet-based algorithm encountering novel lemmas. 1
(Show Context)

Citation Context

... the performance improvement of using CROWN instead of WordNet for measuring semantic similarity when faced with slang or OOV lemmas. Notably, prior semantic similarity benchmarks such as SimLex-999 (=-=Hill et al., 2014-=-) and the ESL test questions (Turney, 2001) have largely omitted these types of words. However, the recent dataset of SemEval-2014 Task 3 (Jurgens et al., 2014) includes similarity judgments between a...

Word Similarity Perception: an Explorative Analysis

by Alice Ruggeri , Luigi Di Caro
"... Abstract Natural language is a medium for expressing things belonging to conceptual and cognitive levels, made of words and grammar rules used to carry semantics. However, its natural ambiguity is the main critical issue that computational systems are generally asked to solve. In this paper, we pro ..."
Abstract - Add to MetaCart
Abstract Natural language is a medium for expressing things belonging to conceptual and cognitive levels, made of words and grammar rules used to carry semantics. However, its natural ambiguity is the main critical issue that computational systems are generally asked to solve. In this paper, we propose to go beyond the current conceptualization of word similarity, i.e., the building block of disambiguation at computational level. First, we analyze the origin of the perceived similarity, studying how conceptual, functional, and syntactic aspects influence its strength. We report the results of a two-stages experiment showing clear similarity perception patterns. Then, based on the insights gained in the cognitive tests, we developed a computational system that automatically predicts word similarity reaching high levels of accuracy.

Specializing Word Embeddings for Similarity or Relatedness

by Douwe Kiela , Felix Hill , Stephen Clark
"... Abstract We demonstrate the advantage of specializing semantic word embeddings for either similarity or relatedness. We compare two variants of retrofitting and a joint-learning approach, and find that all three yield specialized semantic spaces that capture human intuitions regarding similarity an ..."
Abstract - Add to MetaCart
Abstract We demonstrate the advantage of specializing semantic word embeddings for either similarity or relatedness. We compare two variants of retrofitting and a joint-learning approach, and find that all three yield specialized semantic spaces that capture human intuitions regarding similarity and relatedness better than unspecialized spaces. We also show that using specialized spaces in NLP tasks and applications leads to clear improvements, for document classification and synonym selection, which rely on either similarity or relatedness but not both.
(Show Context)

Citation Context

...uch more informative of the topic than knowing that it is a synonym of canine. Conversely, if our embeddings indicate that table is closely related to chair, that does not mean we should translate table into French as chaise. This distinction between “genuine” similarity and associative similarity (i.e., relatedness) is well-known in cognitive science (Tversky, 1977). In NLP, however, semantic spaces are generally evaluated on how well they capture both similarity and relatedness, even though, for many word combinations (such as car and petrol), these two objectives are mutually incompatible (Hill et al., 2014b). In part, this oversight stems from the distributional hypothesis itself: car and petrol do not have the same, or even very similar, meanings, but these two words may well occur in similar contexts. Corpus-driven approaches based on the distributional hypothesis therefore generally learn embeddings that capture both similarity and relatedness reasonably well, but neither perfectly. In this work we demonstrate the advantage of specializing semantic spaces for either similarity or relatedness. Specializing for similarity is achieved by learning from both a corpus and a thesaurus, and for rela...

Published as a conference paper at ICLR 2015 WORD REPRESENTATIONS VIA GAUSSIAN EMBEDDING

by Luke Vilnis, Andrew Mccallum
"... Current work in lexical distributed representations maps each word to a point vector in low-dimensional space. Mapping instead to a density provides many interesting advantages, including better capturing uncertainty about a representa-tion and its relationships, expressing asymmetries more naturall ..."
Abstract - Add to MetaCart
Current work in lexical distributed representations maps each word to a point vector in low-dimensional space. Mapping instead to a density provides many interesting advantages, including better capturing uncertainty about a representa-tion and its relationships, expressing asymmetries more naturally than dot product or cosine similarity, and enabling more expressive parameterization of decision boundaries. This paper advocates for density-based distributed embeddings and presents a method for learning representations in the space of Gaussian distribu-tions. We compare performance on various word embedding benchmarks, inves-tigate the ability of these embeddings to model entailment and other asymmetric relationships, and explore novel properties of the representation. 1
(Show Context)

Citation Context

...t away from its sibling node. 6.4 WORD SIMILARITY BENCHMARKS We evaluate the embeddings on seven different standard word similarity benchmarks (Rubenstein & Goodenough, 1965; Szumlanski et al., 2013; =-=Hill et al., 2014-=-; Miller & Charles, 1991; Bruni et al., 2014; Yang & Powers, 2006; Finkelstein et al., 2001). A comparison to all of the state of the art word-embedding numbers for different dimensionalities as in (B...

Evaluating Learning Language Representations

by Jussi Karlgren, Jimmy Callin, Kevyn Collins-thompson, Amaru Cuba, Ariel Ekgren, David Jurgens, Anna Korhonen, Fredrik Olsson, Magnus Sahlgren, Gavagai Stockholm, Kungl Tekniska Högskolan
"... Abstract. Machine learning offers significant benefits for systems that process and understand natural language: a) lower maintenance and up-keep costs than when using manually-constructed resources, b) easier portability to new domains, tasks, or languages, and c) robust and timely adaptation to si ..."
Abstract - Add to MetaCart
Abstract. Machine learning offers significant benefits for systems that process and understand natural language: a) lower maintenance and up-keep costs than when using manually-constructed resources, b) easier portability to new domains, tasks, or languages, and c) robust and timely adaptation to situation-specific settings. However, the behaviour of an adaptive system is less predictable than when using an edited, stable re-source, which makes quality control a continuous issue. This paper pro-poses an evaluation benchmark for measuring the quality, coverage, and stability of a natural language system as it learns word meaning. Inspired by existing tests for human vocabulary learning, we outline measures for the quality of semantic word representations, such as when learning word embeddings or other distributed representations. These measures high-light differences between the types of underlying learning processes as systems ingest progressively more data.
(Show Context)

Citation Context

... analyses of semantic similarity. The authors explicitly state that their intention was to enable testing on specific and intrinsic characteristics of the testable representations under consideration =-=[1, 5]-=-. While these types of outcome-based tests offer valuable contributions for differentiating the qualities of semantic representations, we propose a different but complementary objective that assesses ...

centre.org

by Author(s Brigadir, Igor Greene, Derek Cunningham, Link To, Igor Brigadir, Derek Greene, Pádraig Cunningham
"... Provided by the author(s) and University College Dublin Library in accordance with publisher policies. Please ..."
Abstract - Add to MetaCart
Provided by the author(s) and University College Dublin Library in accordance with publisher policies. Please
(Show Context)

Citation Context

...ls. Likewise, “oil” and “crude” are synonymous1. Both similarity and relatedness are useful to consider for CDA. It is important to note that DSMs have previously been evaluated with this distinction =-=[16]-=-. Figure 1 shows a toy example of two word spaces. In the space on the left, the word “oil” is represented by a vector [0.50, 0.25], and on the right [0.29, 0.40]. The components of the vectors do not...

Improving Word Representations via Global Visual Context

by Ran Xu, Jiasen Lu, Caiming Xiong, Zhi Yang, Jason J. Corso
"... Visually grounded semantics is a very important aspect in word representation, largely due to its potential to improve many NLP tasks such as information re-trieval, text classification and analysis. We present a new distributed word learn-ing framework which 1) learns word embeddings that better ca ..."
Abstract - Add to MetaCart
Visually grounded semantics is a very important aspect in word representation, largely due to its potential to improve many NLP tasks such as information re-trieval, text classification and analysis. We present a new distributed word learn-ing framework which 1) learns word embeddings that better capture the visually grounded semantics by unifying local document context and global visual context, 2) jointly learns word representation, image representation and language models and 3) focus on better word similarity rather than relatedness. We apply a data set that contains 1 million image-sentence pairs for training and the evaluation on word similarity demonstrates our model outperforms linguistic model without global visual context. 1
(Show Context)

Citation Context

...g to speed up, and also test different window sizes, word vector sizes and global visual contribution factor α. Evaluate Semantics Similarity We use WordSim353 Similarity test set [11] and SimLex-999 =-=[12]-=- to measure the words semantics similarity. For WordSim 353 data set, our vocabulary covers 95% of the annotated ground truth so we only evaluate on the 192 concept pairs. We evaluate the Spearman cor...

ORIGINAL PAPER Cross level semantic similarity: an evaluation framework for universal measures of similarity

by David Jurgens, Mohammad Taher Pilehvar, Roberto Navigli, David Jurgens, Mohammad Taher Pilehvar, Roberto Navigli
"... Abstract Semantic similarity has typically been measured across items of approximately similar sizes. As a result, similarity measures have largely ignored the fact that different types of linguistic item can potentially have similar or even identical meanings, and therefore are designed to compare ..."
Abstract - Add to MetaCart
Abstract Semantic similarity has typically been measured across items of approximately similar sizes. As a result, similarity measures have largely ignored the fact that different types of linguistic item can potentially have similar or even identical meanings, and therefore are designed to compare only one type of lin-guistic item. Furthermore, nearly all current similarity benchmarks within NLP contain pairs of approximately the same size, such as word or sentence pairs, preventing the evaluation of methods that are capable of comparing different sized items. To address this, we introduce a new semantic evaluation called cross-level semantic similarity (CLSS), which measures the degree to which the meaning of a larger linguistic item, such as a paragraph, is captured by a smaller item, such as a sentence. Our pilot CLSS task was presented as part of SemEval-2014, which attracted 19 teams who submitted 38 systems. CLSS data contains a rich mixture of pairs, spanning from paragraphs to word senses to fully evaluate similarity measures that are capable of comparing items of any type. Furthermore, data sources were drawn from diverse corpora beyond just newswire, including domain-specific texts and social media. We describe the annotation process and its challenges, including a comparison with crowdsourcing, and identify the factors that make the dataset a rigorous assessment of a method’s quality. Furthermore, we examine in detail the systems participating in the SemEval task to identify the common factors associated

Not All Neural Embeddings are Born Equal

by Felix Hill, Kyunghyun Cho, Universite ́ De Montréal, Universite ́ De Montréal, Coline Devin, Yoshua Bengio
"... Neural language models learn word representations that capture rich linguistic and conceptual information. Here we investigate the embeddings learned by neural machine translation models. We show that translation-based embeddings outper-form those learned by cutting-edge monolingual models at single ..."
Abstract - Add to MetaCart
Neural language models learn word representations that capture rich linguistic and conceptual information. Here we investigate the embeddings learned by neural machine translation models. We show that translation-based embeddings outper-form those learned by cutting-edge monolingual models at single-language tasks requiring knowledge of conceptual similarity and/or syntactic role. The findings suggest that, while monolingual models learn information about how concepts are related, neural-translation models better capture their true ontological status. It is well known that word representations can be learned from the distributional patterns in corpora. Originally, such representations were constructed by counting word co-occurrences, so that the fea-tures in one word’s representation corresponded to other words [11, 17]. Neural language models, an alternative means to learn word representations, use language data to optimise (latent) features with respect to a language modelling objective. The objective can be to predict either the next word given the initial words of a sentence [4, 14, 8], or simply a nearby word given a single cue word [13, 15]. The representations learned by neural models (sometimes called embeddings) generally outperform those acquired by co-occurrence counting models when applied to NLP tasks [3]. Despite these clear results, it is not well understood how the architecture of neural models affects the information encoded in their embeddings. Here, we explore this question by considering the em-beddings learned by architectures with a very different objective function to monolingual language models: neural machine translation models. We show that translation-based embeddings outperform monolingual embeddings on two types of task: those that require knowledge of conceptual similarity (rather than simply association or relatedness), and those that require knowledge of syntactic role. We discuss what the findings indicate about the information content of different embeddings, and suggest how this content might emerge as a consequence of the translation objective.
(Show Context)

Citation Context

...ng these distances with (gold-standard) human judgements. Table 1 shows the correlations of different model embeddings with three such gold-standard resources, WordSim-353 [1], MEN [5] and SimLex-999 =-=[10]-=-. Interestingly, translation embeddings perform best on SimLex-999, while the two sets of monolingual embeddings perform better on modelling the MEN and WordSim353. To interpret these results, it shou...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University