• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 3,480
Next 10 →

Ensuring quality in crowdsourced search relevance evaluation: The effects of training question distribution

by John Le, Andy Edmonds, Vaughn Hester, Lukas Biewald - In SIGIR 2010 workshop , 2010
"... The use of crowdsourcing platforms like Amazon Mechan-ical Turk for evaluating the relevance of search results has become an effective strategy that yields results quickly and inexpensively. One approach to ensure quality of worker judgments is to include an initial training period and sub-sequent s ..."
Abstract - Cited by 46 (1 self) - Add to MetaCart
The use of crowdsourcing platforms like Amazon Mechan-ical Turk for evaluating the relevance of search results has become an effective strategy that yields results quickly and inexpensively. One approach to ensure quality of worker judgments is to include an initial training period and sub

The use of MMR, diversity-based reranking for reordering documents and producing summaries

by Jaime Carbonell, Jade Goldstein - In SIGIR , 1998
"... jadeQcs.cmu.edu Abstract This paper presents a method for combining query-relevance with information-novelty in the context of text retrieval and summarization. The Maximal Marginal Relevance (MMR) criterion strives to reduce redundancy while maintaining query relevance in re-ranking retrieved docum ..."
Abstract - Cited by 768 (14 self) - Add to MetaCart
systems. However, the clearest advantage is demonstrated in constructing non-redundant multi-document summaries, where MMR results are clearly superior to non-MMR passage selection. 2 Maximal Marginal Relevance Most modem IR search engines produce a ranked list of retrieved documents ordered by declining

Focused crawling: a new approach to topic-specific Web resource discovery

by Soumen Chakrabarti, Martin van den Berg, Byron Dom , 1999
"... The rapid growth of the World-Wide Web poses unprecedented scaling challenges for general-purpose crawlers and search engines. In this paper we describe a new hypertext resource discovery system called a Focused Crawler. The goal of a focused crawler is to selectively seek out pages that are relevan ..."
Abstract - Cited by 637 (10 self) - Add to MetaCart
classifier that evaluates the relevance of a hypertext document with respect to the focus topics, ...

Reexamining the Cluster Hypothesis: Scatter/Gather on Retrieval Results

by Marti A. Hearst, Jan O. Pedersen , 1996
"... We present Scatter/Gather, a cluster-based document browsing method, as an alternative to ranked titles for the organization and viewing of retrieval results. We systematically evaluate Scatter/Gather in this context and find significant improvements over similarity search ranking alone. This resul ..."
Abstract - Cited by 480 (5 self) - Add to MetaCart
We present Scatter/Gather, a cluster-based document browsing method, as an alternative to ranked titles for the organization and viewing of retrieval results. We systematically evaluate Scatter/Gather in this context and find significant improvements over similarity search ranking alone

PAPER Crowdsourcing for Relevance Evaluation

by Omar Alonso, Daniel E. Rose, Benjamin Stewart
"... Relevance evaluation is an essential part of the development and maintenance of information retrieval systems. Yet traditional evaluation approaches have several limitations; in particular, conducting new editorial evaluations of a search system can be very expensive. We describe a new approach to e ..."
Abstract - Add to MetaCart
Relevance evaluation is an essential part of the development and maintenance of information retrieval systems. Yet traditional evaluation approaches have several limitations; in particular, conducting new editorial evaluations of a search system can be very expensive. We describe a new approach

Correlation-based feature selection for machine learning

by Mark A. Hall , 1998
"... A central problem in machine learning is identifying a representative set of features from which to construct a classification model for a particular task. This thesis addresses the problem of feature selection for machine learning through a correlation based approach. The central hypothesis is that ..."
Abstract - Cited by 318 (3 self) - Add to MetaCart
this evaluation formula with an appropriate correlation measure and a heuristic search strategy. CFS was evaluated by experiments on artificial and natural datasets. Three machine learning algorithms were used: C4.5 (a decision tree learner), IB1 (an instance based learner), and naive Bayes. Experiments

Impact of HIT Design on Crowdsourcing Relevance

by Gabriella Kazai, Jaap Kamps, Marijn Koolen, Natasa Milic-frayling
"... In this paper we investigate the design and implementation of effective crowdsourcing tasks in the context of book search evaluation. We observe the impact of aspects of the Human Intelligence Task (HIT) design on the quality of relevance labels provided by the crowd. We assess the output in terms o ..."
Abstract - Add to MetaCart
In this paper we investigate the design and implementation of effective crowdsourcing tasks in the context of book search evaluation. We observe the impact of aspects of the Human Intelligence Task (HIT) design on the quality of relevance labels provided by the crowd. We assess the output in terms

On the evaluation of the quality of relevance assessments collected through crowdsourcing

by Gabriella Kazai, Natasa Milic-frayling - In Proceedings of the SIGIR 2009 Workshop on the Future of IR Evaluation , 2009
"... Established methods for evaluating information retrieval systems rely upon test collections that comprise document corpora, search topics, and relevance assessments. Building large test collections is, however, an expensive and increasingly challenging process. In particular, building a collection w ..."
Abstract - Cited by 10 (0 self) - Add to MetaCart
Established methods for evaluating information retrieval systems rely upon test collections that comprise document corpora, search topics, and relevance assessments. Building large test collections is, however, an expensive and increasingly challenging process. In particular, building a collection

Crowdsourcing document relevance assessment with Mechanical Turk

by Catherine Grady, Matthew Lease - Proc. NAACL HLT Wkshp. Creating Speech and Language Data with Amazon’s Mechanical Turk , 2010
"... We investigate human factors involved in designing effective Human Intelligence Tasks (HITs) for Amazon’s Mechanical Turk1. In particular, we assess document relevance to search queries via MTurk in order to evaluate search engine accuracy. Our study varies four human factors and measures resulting ..."
Abstract - Cited by 35 (3 self) - Add to MetaCart
We investigate human factors involved in designing effective Human Intelligence Tasks (HITs) for Amazon’s Mechanical Turk1. In particular, we assess document relevance to search queries via MTurk in order to evaluate search engine accuracy. Our study varies four human factors and measures resulting

WebMate: A Personal Agent for Browsing and Searching

by Liren Chen, Katia Sycara - In Proceedings of the Second International Conference on Autonomous Agents , 1998
"... The World-Wide Web is developing very fast. Currently, finding useful information on the Web is a time consuming process. In this paper, we present WebMate, an agent that helps users to effectively browse and search the Web. WebMate extends the state of the art in Web-based information retrieval in ..."
Abstract - Cited by 239 (10 self) - Add to MetaCart
provide multiple pages as similarity/relevance guidance for the search. The system extracts and combines relevant keywords from these relevant pages and uses them for keyword refinement. Using these techniques, WebMate provides effective browsing and searching help and also compiles and sends to users
Next 10 →
Results 1 - 10 of 3,480
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University