12 research outputs found
Overview of the CLEF 2017 personalised information retrieval pilot lab (PIR-CLEF 2017)
The Personalised Information Retrieval (PIR-CLEF) Lab
workshop at CLEF 2017 is designed to provide a forum for the exploration of methodologies for the repeatable evaluation of personalised
information retrieval (PIR). The PIR-CLEF 2017 Lab provides a preliminary pilot edition of a Lab task dedicated to personalised search, while
the workshop at the conference is intended to provide a forum for the
discussion of strategies for the evaluation of PIR and extension of the
pilot Lab task. The PIR-CLEF 2017 Pilot Task is the first evaluation
benchmark based on the Cranfield paradigm, with the potential benefits
of producing evaluation results that are easily reproducible. The task is
based on search sessions over a subset of the ClueWeb12 collection, undertaken by 10 users by using a clearly defined and novel methodology.
The collection provides data gathered by the activities undertaken during the search sessions by each participant, including details of relevant
documents as marked by the searchers. The PIR-CLEF 2017 workshop
is intended to review the design and construction of this Pilot collection and to consider the topic of reproducible evaluation of PIR more
generally with the aim of launching a more formal PIR Lab at CLEF
201
Overview of the CLEF 2018 personalised information retrieval lab (PIR-CLEF 2018)
At CLEF 2018, the Personalised Information Retrieval Lab
(PIR-CLEF 2018) has been conceived to provide an initiative aimed at
both providing and critically analysing a new approach to the evaluation
of personalization in Information Retrieval (PIR). PIR-CLEF 2018 is the
first edition of this Lab after the successful Pilot lab organised at CLEF
2017. PIR CLEF 2018 has provided registered participants with the data
sets originally developed for the PIR-CLEF 2017 Pilot task; the data collected are related to real search sessions over a subset of the ClueWeb12
collection, undertaken by 10 users by using a novel methodology. The
data were gathered during the search sessions undertaken by 10 volunteer searchers. Activities during these search sessions included relevance
assessment of a retrieved documents by the searchers. 16 groups registered to participate at PIR-CLEF 2018 and were provided with the data
set to allow them to work on PIR related tasks and to provide feedback
about our proposed PIR evaluation methodology with the aim to create
an effective evaluation task
Overview of the CLEF 2018 personalised information retrieval lab (PIR-CLEF 2018)
At CLEF 2018, the Personalised Information Retrieval Lab
(PIR-CLEF 2018) has been conceived to provide an initiative aimed at
both providing and critically analysing a new approach to the evaluation
of personalization in Information Retrieval (PIR). PIR-CLEF 2018 is the
first edition of this Lab after the successful Pilot lab organised at CLEF
2017. PIR CLEF 2018 has provided registered participants with the data
sets originally developed for the PIR-CLEF 2017 Pilot task; the data collected are related to real search sessions over a subset of the ClueWeb12
collection, undertaken by 10 users by using a novel methodology. The
data were gathered during the search sessions undertaken by 10 volunteer searchers. Activities during these search sessions included relevance
assessment of a retrieved documents by the searchers. 16 groups registered to participate at PIR-CLEF 2018 and were provided with the data
set to allow them to work on PIR related tasks and to provide feedback
about our proposed PIR evaluation methodology with the aim to create
an effective evaluation task
Current Research in Supporting Complex Search Tasks
ABSTRACT ere is broad consensus in the eld of IR that search is complex in many use cases and applications, both on the Web and in domain speci c collections, and both professionally and in our daily life. Yet our understanding of complex search tasks, in comparison to simple look up tasks, is fragmented at best. e workshop addresses many open research questions: What are the obvious use cases and applications of complex search? What are essential features of work tasks and search tasks to take into account? And how do these evolve over time? With a multitude of information, varying from introductory to specialized, and from authoritative to speculative or opinionated, when to show what sources of information? How does the information seeking process evolve and what are relevant di erences between di erent stages? With complex task and search process management, blending searching, browsing, and recommendations, and supporting exploratory search to sensemaking and analytics, UI and UX design pose an overconstrained challenge. How do we evaluate and compare approaches? Which measures should be taken into account? Supporting complex search tasks requires new collaborations across the elds of CHI and IR, and the proposed workshop will bring together a diverse group of researchers to work together on one of the greatest challenges of our eld
A Factored Relevance Model for Contextual Point-of-Interest Recommendation
The challenge of providing personalized and contextually appropriate recommendations to a user is faced in a range of use-cases, e.g., recommendations for movies, places to visit, articles to read etc. In this paper, we focus on one such application, namely that of suggesting 'points of interest' (POIs) to a user given her current location, by leveraging relevant information from her past preferences. An automated contextual recommendation algorithm is likely to work well if it can extract information from the preference history of a user (exploitation) and effectively combine it with information from the user's current context (exploration) to predict an item's 'usefulness' in the new context. To balance this trade-off between exploration and exploitation, we propose a generic unsupervised framework involving a factored relevance model (FRLM), comprising two distinct components, one corresponding to the historical information from past contexts, and the other pertaining to the information from the local context. Our experiments are conducted on the TREC contextual suggestion (TREC-CS) 2016 dataset. The results of our experiments demonstrate the effectiveness of our proposed approach in comparison to a number of standard IR and recommender-based baselines
INEX Tweet Contextualization Task: Evaluation, Results and Lesson Learned
Microblogging platforms such as Twitter are increasingly used for on-line client and market analysis. This motivated the proposal of a new track at CLEF INEX lab of Tweet Contextualization. The objective of this task was to help a user to understand a tweet by providing him with a short explanatory summary (500 words). This summary should be built automatically using resources like Wikipedia and generated by extracting relevant passages and aggregating them into a coherent summary. Running for four years, results show that the best systems combine NLP techniques with more traditional methods. More precisely the best performing systems combine passage retrieval, sentence segmentation and scoring, named entity recognition, text part-of-speech (POS) analysis, anaphora detection, diversity content measure as well as sentence reordering. This paper provides a full summary report on the four-year long task. While yearly overviews focused on system results, in this paper we provide a detailed report on the approaches proposed by the participants and which can be considered as the state of the art for this task. As an important result from the 4 years competition, we also describe the open access resources that have been built and collected. The evaluation measures for automatic summarization designed in DUC or MUC were not appropriate to evaluate tweet contextualization, we explain why and depict in detailed the LogSim measure used to evaluate informativeness of produced contexts or summaries. Finally, we also mention the lessons we learned and that it is worth considering when designing a task