Results 1 -
3 of
3
B.: Towards Data Submissions for Shared Tasks: First Experiences for the Task of Text Alignment
- In: Working Notes Papers of the CLEF 2015 Evaluation Labs. CEUR Workshop Proceedings, CLEF and CEUR-WS.org (Sep 2015), http://www.clef-initiative.eu/publication/ working-notes
"... Abstract This paper reports on the organization of a new kind of shared task that outsources the creation of evaluation resources to its participants. We intro-duce the concept of data submissions for shared tasks, and we use our previous shared task on text alignment as a testbed. A total of eight ..."
Abstract
-
Cited by 3 (1 self)
- Add to MetaCart
Abstract This paper reports on the organization of a new kind of shared task that outsources the creation of evaluation resources to its participants. We intro-duce the concept of data submissions for shared tasks, and we use our previous shared task on text alignment as a testbed. A total of eight evaluation datasets have been submitted by as many participating teams. To validate the submitted datasets, they have been manually peer-reviewed by the participants. Moreover, the submitted datasets have been fed to 31 text alignment approaches in order to learn about the datasets ’ difficulty. The text alignment implementations have been submitted to our shared task in previous years and since been kept operational on the evaluation-as-a-service platform TIRA. 1
D.: Efficient Paragraph based Chunking and Download Filtering for Plagiarism Source Retrieval—Notebook for PAN at CLEF 2015. In: [2
"... Abstract. This paper describes the approach of the system that we built as part of the participation in 'PAN 2015 Source Retrieval' task. Chunking of documents based on paragraphs and efficient download filtering improved the overall performance of the system. Source Retrieval is an impor ..."
Abstract
-
Cited by 1 (0 self)
- Add to MetaCart
(Show Context)
Abstract. This paper describes the approach of the system that we built as part of the participation in 'PAN 2015 Source Retrieval' task. Chunking of documents based on paragraphs and efficient download filtering improved the overall performance of the system. Source Retrieval is an important task of a Plagiarism Detection system
unknown title
"... Abstract This paper presents an overview of the PAN/CLEF evaluation lab. Dur-ing the last decade, PAN has been established as the main forum of text min-ing research focusing on the identification of personal traits of authors left be-hind in texts unintentionally. PAN 2015 comprises three tasks: pl ..."
Abstract
- Add to MetaCart
(Show Context)
Abstract This paper presents an overview of the PAN/CLEF evaluation lab. Dur-ing the last decade, PAN has been established as the main forum of text min-ing research focusing on the identification of personal traits of authors left be-hind in texts unintentionally. PAN 2015 comprises three tasks: plagiarism detec-tion, author identification and author profiling studying important variations of these problems. In plagiarism detection, community-driven corpus construction is introduced as a new way of developing evaluation resources with diversity. In author identification, cross-topic and cross-genre author verification (where the texts of known and unknown authorship do not match in topic and/or genre) is introduced. A new corpus was built for this challenging, yet realistic, task cover-ing four languages. In author profiling, in addition to usual author demographics, such as gender and age, five personality traits are introduced (openness, consci-entiousness, extraversion, agreeableness, and neuroticism) and a new corpus of Twitter messages covering four languages was developed. In total, 53 teams par-ticipated in all three tasks of PAN 2015 and, following the practice of previous editions, software submissions were required and evaluated within the TIRA ex-perimentation framework. 1