• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 1,104
Next 10 →

Integrated NLP Evaluation System for Pluggable Evaluation Metrics with Extensive Interoperable Toolkit

by Yoshinobu Kano, Luke Mccrohon, Sophia Ananiadou
"... To understand the key characteristics of NLP tools, evaluation and comparison against different tools is important. And as NLP applications tend to consist of multiple semiindependent sub-components, it is not always enough to just evaluate complete systems, a fine grained evaluation of underlying c ..."
Abstract - Cited by 3 (1 self) - Add to MetaCart
To understand the key characteristics of NLP tools, evaluation and comparison against different tools is important. And as NLP applications tend to consist of multiple semiindependent sub-components, it is not always enough to just evaluate complete systems, a fine grained evaluation of underlying

On the Predictability of Human Assessment: when Matrix Completion Meets NLP Evaluation

by Guillaume Wisniewski, Université Paris Sud
"... This paper tackles the problem of collecting reliable human assessments. We show that knowing multiple scores for each example instead of a single score results in a more reliable estimation of a system quality. To reduce the cost of collecting these multiple ratings, we propose to use matrix comple ..."
Abstract - Add to MetaCart
This paper tackles the problem of collecting reliable human assessments. We show that knowing multiple scores for each example instead of a single score results in a more reliable estimation of a system quality. To reduce the cost of collecting these multiple ratings, we propose to use matrix completion techniques to predict some scores knowing only scores of other judges and some common ratings. Even if prediction performance is pretty low, decisions made using the predicted score proved to be more reliable than decision based on a single rating of each example. 1

Learning Computational Linguistics through NLP Evaluation Events: the experience of Russian evaluation initiative

by Anastasia Bonch-osmolovskaya, Olga Lyashevskaya, Svetlana Toldova
"... We present in the paper our experience of involving the students of the department of theoretical and computational linguistics of the Moscow State University into full-cycle activities of preparing and evaluating the results of the NLP Evaluation forums, held in 2010 and 2012 in Russia. The forum o ..."
Abstract - Add to MetaCart
We present in the paper our experience of involving the students of the department of theoretical and computational linguistics of the Moscow State University into full-cycle activities of preparing and evaluating the results of the NLP Evaluation forums, held in 2010 and 2012 in Russia. The forum

Evaluating WordNet-based measures of lexical semantic relatedness

by Alexander Budanitsky, Graeme Hirst - Computational Linguistics , 2006
"... The quantification of lexical semantic relatedness has many applications in NLP, and many different measures have been proposed. We evaluate five of these measures, all of which use WordNet as their central resource, by comparing their performance in detecting and correcting real-word spelling error ..."
Abstract - Cited by 321 (0 self) - Add to MetaCart
The quantification of lexical semantic relatedness has many applications in NLP, and many different measures have been proposed. We evaluate five of these measures, all of which use WordNet as their central resource, by comparing their performance in detecting and correcting real-word spelling

NLP Group

by Wim Peters, Yorick Wilks
"... corpus analysis The metonymic relations between word senses is a type of ontological knowledge that is implicitly present in a thesaurus such as WordNet1.6. This paper describes an attempt to make these ontological relations explicit by exploiting the hierarchical structure of WordNet. Resulting pat ..."
Abstract - Add to MetaCart
patterns are evaluated against word sense collocations in Semcor on the assumption that senses that frequently co-occur within documents are an indication of the existence of a lexical relation between them. The results show potential for the extension of the explicit ontological structure of Word

Evaluating Natural Language Processing Systems

by J.R. Galliers, K. Sparck Jones , 1993
"... This report presents a detailed analysis and review of NLP evaluation, in principle and in practice. Part 1 examines evaluation concepts and establishes a framework for NLP system evaluation. This makes use of experience in the related area of information retrieval and the analysis also refers to ev ..."
Abstract - Cited by 148 (0 self) - Add to MetaCart
This report presents a detailed analysis and review of NLP evaluation, in principle and in practice. Part 1 examines evaluation concepts and establishes a framework for NLP system evaluation. This makes use of experience in the related area of information retrieval and the analysis also refers

Automatically Constructing a Dictionary for Information Extraction Tasks

by Ellen Riloff , 1993
"... Knowledge-based natural language processing systems have achieved good success with certain tasks but they are often criticized because they depend on a domain-specific dictionary that requires a great deal of manual knowledge engineering. This knowledge engineering bottleneck makes knowledge-b ..."
Abstract - Cited by 263 (22 self) - Add to MetaCart
-based NLP systems impractical for real-world applications because they cannot be easily scaled up or ported to new domains. In response to this problem, we developed a system called AutoSlog that automatically builds a domain-specific dictionary of concepts for extracting information from text. Using

Towards better nlp system evaluation

by Karen Sparck Jones - In Proceedings of the Human Language Technology Workshop, 102--107 , 1994
"... This paper considers key elements of evaluation methodology, indicating the many points involved and advocating an unpacking approach in specifying an evaluation remit and design. Recognising the importance of both environment variables and system parameters leads to a grid organisation for tests. T ..."
Abstract - Cited by 11 (0 self) - Add to MetaCart
This paper considers key elements of evaluation methodology, indicating the many points involved and advocating an unpacking approach in specifying an evaluation remit and design. Recognising the importance of both environment variables and system parameters leads to a grid organisation for tests

Word representations: A simple and general method for semi-supervised learning

by Joseph Turian, Lev Ratinov, Yoshua Bengio - IN ACL , 2010
"... If we take an existing supervised NLP system, a simple and general way to improve accuracy is to use unsupervised word representations as extra word features. We evaluate Brown clusters, Collobert and Weston (2008) embeddings, and HLBL (Mnih & Hinton, 2009) embeddings of words on both NER and ch ..."
Abstract - Cited by 232 (3 self) - Add to MetaCart
If we take an existing supervised NLP system, a simple and general way to improve accuracy is to use unsupervised word representations as extra word features. We evaluate Brown clusters, Collobert and Weston (2008) embeddings, and HLBL (Mnih & Hinton, 2009) embeddings of words on both NER

Sublanguage Dependent Evaluation: Toward Predicting NLP performances

by Gabriel Illouz , 2000
"... In Natural Language Processing (NLP) Evaluation, such as MUC (Hirshman, 1998), TREC (Harman, 1998), GRACE (Adda et al., 1997), SENSEVAL (Kilgariff, 1998), metrics on the performances, such as precision, recall, or f-measure are used. Nevertheless, performance results are often average measurements c ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
In Natural Language Processing (NLP) Evaluation, such as MUC (Hirshman, 1998), TREC (Harman, 1998), GRACE (Adda et al., 1997), SENSEVAL (Kilgariff, 1998), metrics on the performances, such as precision, recall, or f-measure are used. Nevertheless, performance results are often average measurements
Next 10 →
Results 1 - 10 of 1,104
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University