@MISC{Bando_rmit.edu.au, author = {Lorena Leal Bando and Falk Scholer and Andrew Turpin}, title = {rmit.edu.au}, year = {} }
Share
OpenURL
Abstract
Modern search engines display a summary for each ranked document that is returned in response to a query. These summaries typically include a snippet – a collection of text fragments from the underlying document – that has some relation to the query that is being answered. In this study we investigate how 10 humans construct snippets: participants first generate their own natural language snippet, and then separately extract a snippet by choosing text fragments, for four queries related to two documents. By mapping their generated snippets back to text fragments in the source document using eye tracking data, we observe that participants extract these same pieces of text around 73 % of the time when creating their extractive snippets. In comparison, we notice that automated approaches for extracting snippets only use these same fragments 22 % of the time. However, when the automated methods are evaluated using a position-independent bag-of-words approach, as typically used in the research literature for evaluating snippets, they appear to be much more competitive, with only a 24 point difference in coverage, compared to the human extractive snippets. While there is a 51 point difference when word position is taken into account. In addition to demonstrating this large scope for improvement in snippet generation algorithms with our novel methodology, we also offer a series of observations on the behaviour of participants as they constructed their snippets.