2,492 research outputs found
Pretrained Transformers for Text Ranking: BERT and Beyond
The goal of text ranking is to generate an ordered list of texts retrieved
from a corpus in response to a query. Although the most common formulation of
text ranking is search, instances of the task can also be found in many natural
language processing applications. This survey provides an overview of text
ranking with neural network architectures known as transformers, of which BERT
is the best-known example. The combination of transformers and self-supervised
pretraining has been responsible for a paradigm shift in natural language
processing (NLP), information retrieval (IR), and beyond. In this survey, we
provide a synthesis of existing work as a single point of entry for
practitioners who wish to gain a better understanding of how to apply
transformers to text ranking problems and researchers who wish to pursue work
in this area. We cover a wide range of modern techniques, grouped into two
high-level categories: transformer models that perform reranking in multi-stage
architectures and dense retrieval techniques that perform ranking directly.
There are two themes that pervade our survey: techniques for handling long
documents, beyond typical sentence-by-sentence processing in NLP, and
techniques for addressing the tradeoff between effectiveness (i.e., result
quality) and efficiency (e.g., query latency, model and index size). Although
transformer architectures and pretraining techniques are recent innovations,
many aspects of how they are applied to text ranking are relatively well
understood and represent mature techniques. However, there remain many open
research questions, and thus in addition to laying out the foundations of
pretrained transformers for text ranking, this survey also attempts to
prognosticate where the field is heading
Enhancing the Ranking Context of Dense Retrieval Methods through Reciprocal Nearest Neighbors
Sparse annotation poses persistent challenges to training dense retrieval
models; for example, it distorts the training signal when unlabeled relevant
documents are used spuriously as negatives in contrastive learning. To
alleviate this problem, we introduce evidence-based label smoothing, a novel,
computationally efficient method that prevents penalizing the model for
assigning high relevance to false negatives. To compute the target relevance
distribution over candidate documents within the ranking context of a given
query, we assign a non-zero relevance probability to those candidates most
similar to the ground truth based on the degree of their similarity to the
ground-truth document(s).
To estimate relevance we leverage an improved similarity metric based on
reciprocal nearest neighbors, which can also be used independently to rerank
candidates in post-processing. Through extensive experiments on two large-scale
ad hoc text retrieval datasets, we demonstrate that reciprocal nearest
neighbors can improve the ranking effectiveness of dense retrieval models, both
when used for label smoothing, as well as for reranking. This indicates that by
considering relationships between documents and queries beyond simple geometric
distance we can effectively enhance the ranking context.Comment: EMNLP 202
Neural Methods for Effective, Efficient, and Exposure-Aware Information Retrieval
Neural networks with deep architectures have demonstrated significant
performance improvements in computer vision, speech recognition, and natural
language processing. The challenges in information retrieval (IR), however, are
different from these other application areas. A common form of IR involves
ranking of documents--or short passages--in response to keyword-based queries.
Effective IR systems must deal with query-document vocabulary mismatch problem,
by modeling relationships between different query and document terms and how
they indicate relevance. Models should also consider lexical matches when the
query contains rare terms--such as a person's name or a product model
number--not seen during training, and to avoid retrieving semantically related
but irrelevant results. In many real-life IR tasks, the retrieval involves
extremely large collections--such as the document index of a commercial Web
search engine--containing billions of documents. Efficient IR methods should
take advantage of specialized IR data structures, such as inverted index, to
efficiently retrieve from large collections. Given an information need, the IR
system also mediates how much exposure an information artifact receives by
deciding whether it should be displayed, and where it should be positioned,
among other results. Exposure-aware IR systems may optimize for additional
objectives, besides relevance, such as parity of exposure for retrieved items
and content publishers. In this thesis, we present novel neural architectures
and methods motivated by the specific needs and challenges of IR tasks.Comment: PhD thesis, Univ College London (2020
- …