lecture notes in computer science 9283978-3-319-24027...lecture notes in computer science 9283...

20
Lecture Notes in Computer Science 9283 Commenced Publication in 1973 Founding and Former Series Editors: Gerhard Goos, Juris Hartmanis, and Jan van Leeuwen Editorial Board David Hutchison Lancaster University, Lancaster, UK Takeo Kanade Carnegie Mellon University, Pittsburgh, PA, USA Josef Kittler University of Surrey, Guildford, UK Jon M. Kleinberg Cornell University, Ithaca, NY, USA Friedemann Mattern ETH Zurich, Zürich, Switzerland John C. Mitchell Stanford University, Stanford, CA, USA Moni Naor Weizmann Institute of Science, Rehovot, Israel C. Pandu Rangan Indian Institute of Technology, Madras, India Bernhard Steffen TU Dortmund University, Dortmund, Germany Demetri Terzopoulos University of California, Los Angeles, CA, USA Doug Tygar University of California, Berkeley, CA, USA Gerhard Weikum Max Planck Institute for Informatics, Saarbrücken, Germany

Upload: others

Post on 05-Jul-2020

10 views

Category:

Documents


0 download

TRANSCRIPT

Lecture Notes in Computer Science 9283

Commenced Publication in 1973Founding and Former Series Editors:Gerhard Goos, Juris Hartmanis, and Jan van Leeuwen

Editorial Board

David HutchisonLancaster University, Lancaster, UK

Takeo KanadeCarnegie Mellon University, Pittsburgh, PA, USA

Josef KittlerUniversity of Surrey, Guildford, UK

Jon M. KleinbergCornell University, Ithaca, NY, USA

Friedemann MatternETH Zurich, Zürich, Switzerland

John C. MitchellStanford University, Stanford, CA, USA

Moni NaorWeizmann Institute of Science, Rehovot, Israel

C. Pandu RanganIndian Institute of Technology, Madras, India

Bernhard SteffenTU Dortmund University, Dortmund, Germany

Demetri TerzopoulosUniversity of California, Los Angeles, CA, USA

Doug TygarUniversity of California, Berkeley, CA, USA

Gerhard WeikumMax Planck Institute for Informatics, Saarbrücken, Germany

More information about this series at http://www.springer.com/series/7409

Josiane Mothe • Jacques SavoyJaap Kamps • Karen Pinel-SauvagnatGareth J.F. Jones • Eric SanJuanLinda Cappellato • Nicola Ferro (Eds.)

Experimental IR MeetsMultilinguality,Multimodality, and Interaction6th International Conferenceof the CLEF Association, CLEF’15Toulouse, France, September 8–11, 2015Proceedings

123

EditorsJosiane MotheInstitut de Recherche en Informatique

de ToulouseToulouseFrance

Jacques SavoyUniversity of NeuchâtelNeuchâtelSwitzerland

Jaap KampsUniversity of AmsterdamAmsterdamThe Netherlands

Karen Pinel-SauvagnatInstitut de Recherche en Informatique

de ToulouseToulouseFrance

Gareth J.F. JonesDublin City UniversityDublinIreland

Eric SanJuanUniversité d’Avignon et des Pays

de VaucluseAvignonFrance

Linda CappellatoUniversity of PaduaPaduaItaly

Nicola FerroUniversity of PaduaPaduaItaly

ISSN 0302-9743 ISSN 1611-3349 (electronic)Lecture Notes in Computer ScienceISBN 978-3-319-24026-8 ISBN 978-3-319-24027-5 (eBook)DOI 10.1007/978-3-319-24027-5

Library of Congress Control Number: 2015947945

LNCS Sublibrary: SL3 – Information Systems and Applications, incl. Internet/Web, and HCI

Springer Cham Heidelberg New York Dordrecht London© Springer International Publishing Switzerland 2015This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part of thematerial is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation,broadcasting, reproduction on microfilms or in any other physical way, and transmission or informationstorage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology nowknown or hereafter developed.The use of general descriptive names, registered names, trademarks, service marks, etc. in this publicationdoes not imply, even in the absence of a specific statement, that such names are exempt from the relevantprotective laws and regulations and therefore free for general use.The publisher, the authors and the editors are safe to assume that the advice and information in this book arebelieved to be true and accurate at the date of publication. Neither the publisher nor the authors or the editorsgive a warranty, express or implied, with respect to the material contained herein or for any errors oromissions that may have been made.

Printed on acid-free paper

Springer International Publishing AG Switzerland is part of Springer Science+Business Media(www.springer.com)

Preface

Since 2000, the Conference and Labs of the Evaluation Forum (CLEF) has played aleading role in stimulating research and innovation in the domain of multimodal andmultilingual information access. Initially founded as the Cross-Language EvaluationForum and running in conjunction with the European Conference on Digital Libraries(ECDL/TPDL), CLEF became a standalone event in 2010 combining a peer-reviewedconference with a multi-track evaluation forum. CLEF 20151 was hosted by the Institutde Recherche en Informatique de Toulouse UMR 5505 CNRS, Université de Toulouse,France, SIG team.

The CLEF conference addresses all aspects of information access in any modalityand language. The conference has a clear focus on experimental IR as done at eval-uation forums (CLEF Labs, TREC, NTCIR, FIRE, MediaEval, RomIP, SemEval,TAC, …), paying special attention to the challenges of multimodality, multilinguality,and interactive search. We invited submissions on significant new insights demon-strated on the resulting IR test collections, on analysis of IR test collections andevaluation measures, as well as on concrete proposals to push the boundaries of theCranfield/TREC/CLEF paradigm. The conference format consisted of keynotes, con-tributed papers, lab sessions, and poster sessions, including reports from otherbenchmarking initiatives from around the world. It was an honor and a privilege tohave Gregory Grefenstette (INRIA Saclay, France), Mounia Lalmas (Yahoo Labs,London, UK), and Douglas W. Oard (University of Maryland, USA) as keynotespeakers. Greg talked about personal information systems and personal semantics,Mounia addressed the topic of user engagement evaluation, and Doug examined issuesin privacy and ethics when searching among secrets.

CLEF 2015 received a total of 68 submissions, a dramatic increase over previousyears. Each submission was reviewed by at least three PC members, and the twoprogram chairs oversaw the reviewing and often extensive follow-up discussion. Wherethe discussion was not sufficient to make a decision, the paper, went through an extrareview by the PC. A novel feature of the CLEF 2015 conference was to invite CLEF2014 lab organizers to nominate a “best of the labs” paper, which was reviewed as afull paper submission to the CLEF 2015 conference according to the same reviewcriteria and PC. This resulted in 8 full papers accepted corresponding to each to theCLEF 2014 labs. We received 24 regular full paper submissions, of which 8 (33 %) fullpapers were accepted for regular oral presentation, 7 further full paper submissions (29 %,making a total of 63 %) accepted with short oral presentation and a poster. We received 36short paper submissions, and accepted 20 (55 %).

1 http://clef2015.clef-initiative.eu/

In addition to these talks, the eight benchmarking labs reported results of their yearlong activities in overview talks and lab sessions2. The eight labs running as part ofCLEF 2015 were as follows:

CLEFeHealth provided scenarios aiming to ease patients’ and nurses’ understand-ing and accessing of eHealth information. The goals of the lab were to develop pro-cessing methods and resources in a multilingual setting, to enrich difficult-to-understandeHealth texts, and provide valuable documentation. The tasks were: informationextraction from clinical data, and user-centered health information retrieval.

ImageCLEF provided four main tasks with a global objective of benchmarkingautomatic annotation and indexing of images. The tasks tackled different aspects of theannotation problem and aimed at supporting and promoting cutting-edge researchaddressing the key challenges in the field: image annotation, medical classification,medical clustering, and liver CT annotation.

LifeCLEF provided image-based plant, bird, and fish identification tasks addressingmultimodal data by (i) considering birds and fish in addition to plants, (ii) consideringaudio and video content in addition to images, (iii) scaling-up the evaluation data tohundreds of thousands of life media records and thousands of living species. The taskswere: an audio record-based bird identification task (BirdCLEF), an image-based plantidentification task (PlantCLEF), and a fish video surveillance task (FishCLEF).

Living Labs for IR (LL4IR) provided a benchmarking platform for researchers toevaluate their ranking systems in a live setting with real users in their natural task envi-ronments. The lab acted as a proxy between commercial organizations (live environments)and lab participants (experimental systems), facilitated data exchange, and made com-parisons between the participating systems. The task was: product search and web search.

News Recommendation Evaluation Lab (NEWSREEL) provided two tasksdesigned to address the challenge of real-time news recommendation. Participantscould: a) develop news recommendation algorithms and b) have them tested by mil-lions of users over the period of a few weeks in a living lab. The tasks were: benchmarknews recommendations in a living lab, benchmarking news recommendations in asimulated environment.

Uncovering Plagiarism, Authorship, and Social Software Misuse (PAN) pro-vided evaluation of uncovering plagiarism, authorship, and social software misuse.PAN offered three tasks at CLEF 2015 with new evaluation resources consisting oflarge-scale corpora, performance measures, and web services that allowed for mean-ingful evaluations. The main goal was to provide for sustainable and reproducibleevaluations, to get a clear view of the capabilities of state-of-the-art algorithms. Thetasks were: plagiarism detection, author identification, and author profiling.

Question Answering (QA) provided QA from the starting point of a natural lan-guage question. However, answering some questions may need to query linked data

2 The full details for each lab are contained in a separate publication, the Working Notes, which areavailable online at http://ceur-ws.org/Vol-1391/.

VI Preface

(especially if aggregations or logical inferences are required); whereas some questionsmay need textual inferences and querying free-text. Answering some queries may needboth. The tasks were: QALD: Question Answering over Linked Data, entrance exams:questions from reading tests, BioASQ: large-scale biomedical semantic indexing, andBioASQ: biomedical question answering.

Social Book Search (SBS) provided evaluation of real-world information needswhich are generally complex, yet almost all research focuses instead on either relativelysimple search based on queries or recommendation based on profiles. The goal of theSocial Book Search Lab was to investigate techniques to support users in complexbook search tasks that involve more than just a query and results list. The tasks were:suggestion track, and interactive track.

A rich social program was organized in conjunction with the conference, startingwith a welcome reception with local food and wine specialities, continuing with a cityhall reception, which included the local band “La mal Coiffée”. The social dinner wasenjoyed in a famous organic restaurant named “Saveur Bio”, and master classes in(1) traditional polyphonic singing with Bastien Zaoui from the famous Vox Bigerriband and (2) wine and food pairing with Yves Cinotti, were also offered.

The success of CLEF 2015 would not have been possible without the huge effort ofseveral people and organizations, including the CLEF Association3, the ProgramCommittee, the Lab Organizing Committee, the Local Organization Committee inToulouse, the reviewers, and the many students and volunteers who contributed alongthe way. We would like to acknowledge the Institut de Recherche en Informatique deToulouse UMR 5505 CNRS and its director, Prof. Michel Daydé, for the support we got,first for bidding to host the conference, then for organizing it. We also received thesupport from the following universities and schools: Ecole supérieure du professorat etde l’éducation, Université Toulouse-Jean Jaurès, Université Paul Sabatier, and Universitédu Capitole. We also gratefully acknowledge the support we received from our sponsors.The ESF Research Networking Program ELIAS, the ACM SIG-IR, the UniversitéToulouse-Jean Jaurès, and the Région Midi-Pyrénées for their strong financial support;but also: Springer, the Université Paul Sabatier, Institut de Recherche en Informatiquede Toulouse UMR 5505 CNRS, INFORSID, Université Toulouse Capitole, EGC,ARIA, and ACL. The level of sponsorship allowed us to offer 20 grants for students inaddition to a free registration for the 25 volunteers including 11 further students.

July 2015 Josiane MotheJacques Savoy

Jaap KampsKaren Pinel-Sauvagnat

Gareth J.F. JonesEric SanJuan

Linda CappellatoNicola Ferro

3 http://www.clef-initiative.eu/association

Preface VII

Organization

CLEF 2015 Conference and Labs of the Evaluation Forum, Experimental IR MeetsMultilinguality, Multimodality, and Interaction, was organized by the University ofToulouse, France.

General Chair

Josiane Mothe IRIT, Université de Toulouse, FranceJacques Savoy University of Neuchâtel, Switzerland

Program Chair

Jaap Kamps University of Amsterdam, The NetherlandsKaren Pinel-Sauvagnat IRIT, Université de Toulouse, France

Lab Chair

Gareth J.F. Jones Dublin City University, IrelandEric SanJuan Université d’Avignon et des Pays du Vaucluse, France

Program Committee

Maristella Agosti University of Padua, ItalyKrisztian Balog University of Stavanger, NorwayPatrice Bellot LSIS - Université de Marseille, FranceToine Bogers Aalborg University Copenhagen, DenmarkMohand Boughanem IRIT - Université Paul Sabatier Toulouse 3, FranceGuillaume Cabanac IRIT - Université Paul Sabatier Toulouse 3, FranceTiziana Catarci Università di Roma “La Sapienza”, ItalyPaul Clough University of Sheffield, UKNicola Ferro University of Padua, ItalyNorbert Fuhr University of Duisburg-Essen, GermanyEric Gaussier Université Joseph Fourier (Grenoble I), FranceLorraine Goeuriot Université Joseph Fourier (Grenoble I), FranceJulio Gonzalo UNED, Madrid, SpainAllan Hanbury Vienna University of Technology, AustriaDonna Harman NIST, USADjoerd Hiemstra University of Twente, The NetherlandsFrank Hopfgartner University of Glasgow, UKGilles Hubert IRIT - Université Paul Sabatier Toulouse 3, FrancePeter Ingwersen University of Copenhagen, DenmarkAlexis Joly INRIA Sophia-Antipolis, FranceGareth J.F. Jones Dublin City University, Ireland

Evangelos Kanoulas University of Amsterdam, The NetherlandsGabriella Kazai Lumi, UKJaana Kekäläinen University of Tampere, FinlandLiadh Kelly Trinity College Dublin, IrelandBenjamin Kille DAI Lab, Berlin Institute of Technology, GermanyMarijn Koolen University of Amsterdam, The NetherlandsBirger Larsen Aalborg University, DenmarkMihai Lupu Vienna University of Technology, AustriaThomas Mandl University of Hildesheim, GermanyHenning Müller HES-SO, University of Applied Sciences Western

Switzerland, SwitzerlandJian-Yun Ni Université de Montréal, CanadaIadh Ounis University of Glasgow, UKGabriella Pasi Università degli Studi di Milano Bicocca, ItalyAnselmo Peñas NLP and IR Group, UNED, SpainBenjamin Piwowarski CNRS/Université Pierre et Marie Curie, FranceMartin Potthast Bauhaus University Weimar, GermanyPaolo Rosso Technical University of Valencia, SpainEric SanJuan Université d’Avignon, FranceRalf Schenkel Universität Passau, GermanyAnne Schuth University of Amsterdam, The NetherlandsEfstathios Stamatatos University of the Aegean, GreeceBenno Stein Bauhaus-Universität Weimar, GermanyLynda Tamine IRIT - Université Paul Sabatier Toulouse 3, FranceXavier Tannier LIMSI-CNRS, Université Paris-Sud, FranceTheodora Tsikrika Information Technologies Institute, CERTH, GreeceChristina Unger CITEC, Universität Bielefeld, GermanyMauricio Villegas Universitat Politècnica de València, Spain

Local Organization

Adrian Chifu IRIT, Université de Toulouse, France (Sponsoring)Véronique Debats IRIT, France (Communication)Marlène Giamporcaro SAIC, INP Toulouse, France (Administration and

Registration)Laure Soulier IRIT, Université de Toulouse, France (Advertising)Nathalie

Valles-ParlengeauIRIT, Université de Toulouse, France (Co-resp.

for UT1-Capitole University)

X Organization

Platinum Sponsors

Silver Sponsors

Organization XI

Bronze Sponsors

XII Organization

Keynotes

Personal Information Systemsand Personal Semantics

Gregory Grefenstette

INRIA Saclay, France

People generally think of Big Data as something generated by machines or largecommunities of people interacting with the digital world. But technological progressmeans that each individual is currently, or soon will be, generating masses of digitaldata in their everyday lives. In every interaction with an application, every web pagevisited, every time your telephone is turned on, you generate information aboutyourself, Personal Big Data. With the rising adoption of quantified self gadgets, and theforeseeable adoption of intelligent glasses capturing daily life, the quantity of personalBig Data will only grow. In this Personal Big Data, as in other Big Data, a key problemis aligning concepts in the same semantic space. While concept alignment in the publicsphere is an understood, though unresolved, problem, what does ontological organi-zation of a personal space look like? Is it idiosyncratic, or something that can be sharedbetween people? We will describe our current approach to this problem of organizingpersonal data and creating and exploiting a personal semantics.

Evaluating the Search Experience:From Retrieval Effectiveness to User Engagement

Mounia Lalmas

Yahoo Labs, London, UK

Building retrieval systems that return results to users that satisfy their information needis one thing; Information Retrieval has a long history in evaluating how effectiveretrieval systems are. Many evaluation initiatives such as TREC and CLEF haveallowed organizations worldwide to evaluate and compare retrieval approaches.Building a retrieval system that not only returns good results to users, but does so in away that users will want to use that system again is something more challenging; apositive search experience has been shown to lead to users engaging long-term with theretrieval system. In this talk, I will review state-of-the-art approaches concerned withevaluating retrieval effectiveness. I will then focus on those approaches aiming atevaluating user engagement, and describe current works in this area. The talk will endwith the proposal of a framework incorporating effectiveness evaluation into userengagement. An important component of this framework is to consider both within-and across-search session measurement.

Beyond Information Retrieval:When and How Not to Find Things

Douglas W. Oard

University of Maryland, USA

The traditional role of a search engine is much like the traditional role of a library:generally the objective is to help people find things. As we get better at this, however,we have been encountering an increasing number of cases in which some things that weknow exist simply should not be found. Some well known examples include removal ofimproperly posted copyrighted material from search engine indexes, and the evolvinglegal doctrine that is now commonly referred to as the “right to be forgotten.” Somesuch cases are simple, relying on users to detect specific content that should be flushedfrom a specific index. Other cases, however, are more complex. For example, in theaspect of the civil litigation process known as e-discovery, one side may be entitled towithhold entire classes of material that may not have been labeled in advance (becauseof attorney-client privilege). An even more complex example is government trans-parency, in which for public policy reasons we may want to make some informationpublic, despite that information being intermixed with other information that must beprotected. Professional archivists have long dealt with such challenges, so perhaps weshould start thinking about how to build search engines that act less like a library andmore like an archive. In this talk, I will use these and other examples to introduce theidea of “search among secrets” in which the goal is to help some users find somecontent while protecting some content from some users (or some uses). We’ll divedown to look at how this actually works today in a few specific cases, with particularattention to how queries are formulated and which parts of the process are, or might be,automated. With that as background, I will then offer a few initial thoughts on how wemight evaluate such systems. I’ll conclude with an invitation to think together abouthow information retrieval researchers might, together with others, begin to tackle thesechallenges.

Contents

Experimental IR

Experimental Study on Semi-structured Peer-to-Peer Information RetrievalNetwork. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

Rami S. Alkhawaldeh and Joemon M. Jose

Evaluating Stacked Marginalised Denoising Autoencoders Within DomainAdaptation Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

Boris Chidlovskii, Gabriela Csurka, and Stephane Clinchant

Language Variety Identification Using Distributed Representationsof Words and Documents. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

Marc Franco-Salvador, Francisco Rangel, Paolo Rosso, Mariona Taulé,and M. Antònia Martít

Evaluating User Image Tagging Credibility . . . . . . . . . . . . . . . . . . . . . . . . 41Alexandru Lucian Ginsca, Adrian Popescu, Mihai Lupu, Adrian Iftene,and Ioannis Kanellos

Web and Social Media

Tweet Expansion Method for Filtering Task in Twitter . . . . . . . . . . . . . . . . 55Payam Karisani, Farhad Oroumchian, and Maseud Rahgozar

Real-Time Entity-Based Event Detection for Twitter . . . . . . . . . . . . . . . . . . 65Andrew J. McMinn and Joemon M. Jose

A Comparative Study of Click Models for Web Search . . . . . . . . . . . . . . . . 78Artem Grotov, Aleksandr Chuklin, Ilya Markov, Luka Stout,Finde Xumara, and Maarten de Rijke

Evaluation of Pseudo Relevance Feedback Techniques for Cross VerticalAggregated Search . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91

Hermann Ziak and Roman Kern

Long Papers with Short Presentation

Analysing the Role of Representation Choices in Portuguese RelationExtraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105

Sandra Collovini, Marcelo de Bairros P. Filho, and Renata Vieira

An Investigation of Cross-Language Information Retrievalfor User-Generated Internet Video. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117

Ahmad Khwileh, Debasis Ganguly, and Gareth J.F. Jones

Benchmark of Rule-Based Classifiers in the News Recommendation Task . . . 130Tomáš Kliegr and Jaroslav Kuchař

Enhancing Medical Information Retrieval by Exploiting a Content-BasedRecommender Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142

Wei Li and Gareth J.F. Jones

Summarizing Citation Contexts of Scientific Publications . . . . . . . . . . . . . . . 154Sandra Mitrović and Henning Müller

A Multiple-Stage Approach to Re-ranking Medical Documents . . . . . . . . . . . 166Heung-Seon Oh, Yuchul Jung, and Kwang-Young Kim

Exploring Behavioral Dimensions in Session Effectiveness. . . . . . . . . . . . . . 178Teemu Pääkkönen, Kalervo Järvelin, Jaana Kekäläinen,Heikki Keskustalo, Feza Baskaya, David Maxwell, and Leif Azzopardi

Short Papers

META TEXT ALIGNER: Text Alignment Based on Predicted PlagiarismRelation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193

Samira Abnar, Mostafa Dehghani, and Azadeh Shakery

Automatic Indexing of Journal Abstracts with Latent Semantic Analysis . . . . 200Joel Robert Adams and Steven Bedrick

Shadow Answers as an Intermediary in Email Answer Retrieval . . . . . . . . . . 209Alyaa Alfalahi, Gunnar Eriksson, and Eriks Sneiders

Are Topically Diverse Documents Also Interesting? . . . . . . . . . . . . . . . . . . 215Hosein Azarbonyad, Ferron Saan, Mostafa Dehghani, Maarten Marx,and Jaap Kamps

Modeling of the Question Answering Task in the YodaQA System . . . . . . . . 222Petr Baudiš and Jan Šedivý

Unfair Means: Use Cases Beyond Plagiarism . . . . . . . . . . . . . . . . . . . . . . . 229Paul Clough, Peter Willett, and Jessie Lim

Instance-Based Learning for Tweet Monitoring and Categorization . . . . . . . . 235Julien Gobeill, Arnaud Gaudinat, and Patrick Ruch

XX Contents

Are Test Collections “Real”? Mirroring Real-World Complexityin IR Test Collections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 241

Melanie Imhof and Martin Braschler

Evaluation of Manual Query Expansion Rules on a Domain SpecificFAQ Collection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 248

Mladen Karan and Jan Šnajder

Evaluating Learning Language Representations . . . . . . . . . . . . . . . . . . . . . . 254Jussi Karlgren, Jimmy Callin, Kevyn Collins-Thompson,Amaru Cuba Gyllensten, Ariel Ekgren, David Jurgens, Anna Korhonen,Fredrik Olsson, Magnus Sahlgren, and Hinrich Schütze

Automatic Segmentation and Deep Learning of Bird Sounds . . . . . . . . . . . . 261Hendrik Vincent Koops, Jan van Balen, and Frans Wiering

The Impact of Noise in Web Genre Identification . . . . . . . . . . . . . . . . . . . . 268Dimitrios Pritsos and Efstathios Stamatatos

On the Multilingual and Genre Robustness of EmoGraphs for AuthorProfiling in Social Media . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 274

Francisco Rangel and Paolo Rosso

Is Concept Mapping Useful for Biomedical Information Retrieval? . . . . . . . . 281Wei Shen and Jian-Yun Nie

Using Health Statistics to Improve Medical and Health Search . . . . . . . . . . . 287Tawan Sierek and Allan Hanbury

Determining Window Size from Plagiarism Corpus for Stylometric Features . . . 293Šimon Suchomel and Michal Brandejs

Effect of Log-Based Query Term Expansion on Retrieval Effectivenessin Patent Searching . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300

Wolfgang Tannebaum, Parvaz Mahdabi, and Andreas Rauber

Integrating Mixed-Methods for Evaluating Information Access Systems . . . . . 306Simon Wakeling and Paul Clough

Teaching the IR Process Using Real Experiments Supported by GameMechanics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 312

Thomas Wilhelm-Stein and Maximilian Eibl

Tweet Contextualization Using Association Rules Mining and DBpedia. . . . . 318Meriem Amina Zingla, Chiraz Latiri, and Yahya Slimani

Contents XXI

Best of the Labs

Search-Based Image Annotation: Extracting Semantics from SimilarImages. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 327

Petra Budikova, Michal Batko, Jan Botorek, and Pavel Zezula

NLP-Based Classifiers to Generalize Expert Assessments in E-Reputation . . . 340Jean-Valère Cossu, Emmanuel Ferreira, Killian Janod, Julien Gaillard,and Marc El-Bèze

A Method for Short Message Contextualization: Experimentsat CLEF/INEX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 352

Liana Ermakova

Towards Automatic Large-Scale Identification of Birds in AudioRecordings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364

Mario Lasseck

Optimizing and Evaluating Stream-Based News RecommendationAlgorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376

Andreas Lommatzsch and Sebastian Werner

Information Extraction from Clinical Documents: TowardsDisease/Disorder Template Filling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 389

Veera Raghavendra Chikka, Nestor Mariyasagayam, Yoshiki Niwa,and Kamalakar Karlapalem

Adaptive Algorithm for Plagiarism Detection: The Best-PerformingApproach at PAN 2014 Text Alignment Competition . . . . . . . . . . . . . . . . . 402

Miguel A. Sanchez-Perez, Alexander Gelbukh, and Grigori Sidorov

Question Answering via Phrasal Semantic Parsing. . . . . . . . . . . . . . . . . . . . 414Kun Xu, Yansong Feng, Songfang Huang, and Dongyan Zhao

Labs Overviews

Overview of the CLEF eHealth Evaluation Lab 2015 . . . . . . . . . . . . . . . . . 429Lorraine Goeuriot, Liadh Kelly, Hanna Suominen, Leif Hanlen,Aurélie Névéol, Cyril Grouin, João Palotti, and Guido Zuccon

General Overview of ImageCLEF at the CLEF 2015 Labs . . . . . . . . . . . . . . 444Mauricio Villegas, Henning Müller, Andrew Gilbert, Luca Piras,Josiah Wang, Krystian Mikolajczyk, Alba G. Seco de Herrera,Stefano Bromuri, M. Ashraful Amin, Mahmood Kazi Mohammed,Burak Acar, Suzan Uskudarli, Neda B. Marvasti, José F. Aldana,and María del Mar Roldán García

XXII Contents

LifeCLEF 2015: Multimedia Life Species Identification Challenges. . . . . . . . 462Alexis Joly, Hervé Goëau, Hervé Glotin, Concetto Spampinato,Pierre Bonnet, Willem-Pier Vellinga, Robert Planqué, Andreas Rauber,Simone Palazzo, Bob Fisher, and Henning Müller

Overview of the Living Labs for Information Retrieval Evaluation (LL4IR)CLEF Lab 2015 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 484

Anne Schuth, Krisztian Balog, and Liadh Kelly

Stream-Based Recommendations: Online and Offline Evaluationas a Service . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 497

Benjamin Kille, Andreas Lommatzsch, Roberto Turrin, András Serény,Martha Larson, Torben Brodt, Jonas Seiler, and Frank Hopfgartner

Overview of the PAN/CLEF 2015 Evaluation Lab . . . . . . . . . . . . . . . . . . . 518Efstathios Stamatatos, Martin Potthast, Francisco Rangel, Paolo Rosso,and Benno Stein

Overview of the CLEF Question Answering Track 2015 . . . . . . . . . . . . . . . 539Anselmo Peñas, Christina Unger, Georgios Paliouras,and Ioannis Kakadiaris

Overview of the CLEF 2015 Social Book Search Lab . . . . . . . . . . . . . . . . . 545Marijn Koolen, Toine Bogers, Maria Gäde, Mark Hall,Hugo Huurdeman, Jaap Kamps, Mette Skov, Elaine Toms,and David Walsh

Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 565

Contents XXIII