9 research outputs found

    SemRank: ranking refinement strategy by using the semantic intensity

    Get PDF
    AbstractThe ubiquity of the multimedia has raised a need for the system that can store, manage, structured the multimedia data in such a way that it can be retrieved intelligently. One of the current issues in media management or data mining research is ranking of retrieved documents. Ranking is one of the provocative problems for information retrieval systems. Given a user query comes up with the millions of relevant results but if the ranking function cannot rank it according to the relevancy than all results are just obsolete. However, the current ranking techniques are in the level of keyword matching. The ranking among the results is usually done by using the term frequency. This paper is concerned with ranking the document relying merely on the rich semantic inside the document instead of the contents. Our proposed ranking refinement strategy known as SemRank, rank the document based on the semantic intensity. Our approach has been applied on the open benchmark LabelMe dataset and compared against one of the well known ranking model i.e. Vector Space Model (VSM). The experimental results depicts that our approach has achieved significant improvement in retrieval performance over the state of the art ranking methods

    Intent-aware search result diversification

    Full text link
    Search result diversification has gained momentum as a way to tackle ambiguous queries. An effective approach to this problem is to explicitly model the possible aspects underlying a query, in order to maximise the estimated relevance of the retrieved documents with respect to the different aspects. However, such aspects themselves may represent information needs with rather distinct intents (e.g., informational or navigational). Hence, a diverse ranking could benefit from applying intent-aware retrieval models when estimating the relevance of documents to different aspects. In this paper, we propose to diversify the results retrieved for a given query, by learning the appropriateness of different retrieval models for each of the aspects underlying this query. Thorough experiments within the evaluation framework provided by the diversity task of the TREC 2009 and 2010 Web tracks show that the proposed approach can significantly improve state-of-the-art diversification approaches

    A cross-benchmark comparison of 87 learning to rank methods

    Get PDF
    Learning to rank is an increasingly important scientific field that comprises the use of machine learning for the ranking task. New learning to rank methods are generally evaluated on benchmark test collections. However, comparison of learning to rank methods based on evaluation results is hindered by the absence of a standard set of evaluation benchmark collections. In this paper we propose a way to compare learning to rank methods based on a sparse set of evaluation results on a set of benchmark datasets. Our comparison methodology consists of two components: (1) Normalized Winning Number, which gives insight in the ranking accuracy of the learning to rank method, and (2) Ideal Winning Number, which gives insight in the degree of certainty concerning its ranking accuracy. Evaluation results of 87 learning to rank methods on 20 well-known benchmark datasets are collected through a structured literature search. ListNet, SmoothRank, FenchelRank, FSMRank, LRUF and LARF are Pareto optimal learning to rank methods in the Normalized Winning Number and Ideal Winning Number dimensions, listed in increasing order of Normalized Winning Number and decreasing order of Ideal Winning Number

    Expansion sĂ©lective de requĂȘtes par apprentissage

    Get PDF
    Si l’expansion de requĂȘte automatique amĂ©liore en moyenne la qualitĂ© de recherche, elle peut la dĂ©grader pour certaines requĂȘtes. Ainsi, certains travaux s’intĂ©ressent Ă  dĂ©velopper des approches sĂ©lectives qui choisissent la fonction de recherche ou d’expansion en fonction des requĂȘtes. La plupart des approches sĂ©lectives utilisent un processus d’apprentissage sur des caractĂ©ristiques de requĂȘtes passĂ©es et sur les performances obtenues. Cet article prĂ©sente une nouvelle mĂ©thode d’expansion sĂ©lective qui se base sur des prĂ©dicteurs de difficultĂ© des requĂȘtes, prĂ©dicteurs linguistiques et statistiques. Le modĂšle de dĂ©cision est appris par un SVM. Nous montrons l’efficacitĂ© de la mĂ©thode sur des collections TREC standards. Les modĂšles appris ont classĂ© les requĂȘtes de test avec plus de 90% d’exactitude. Par ailleurs, la MAP est amĂ©liorĂ©e de plus de 11%, comparĂ©e Ă  des mĂ©thodes non sĂ©lectives

    Learning to select for information retrieval

    Get PDF
    The effective ranking of documents in search engines is based on various document features, such as the frequency of the query terms in each document, the length, or the authoritativeness of each document. In order to obtain a better retrieval performance, instead of using a single or a few features, there is a growing trend to create a ranking function by applying a learning to rank technique on a large set of features. Learning to rank techniques aim to generate an effective document ranking function by combining a large number of document features. Different ranking functions can be generated by using different learning to rank techniques or on different document feature sets. While the generated ranking function may be uniformly applied to all queries, several studies have shown that different ranking functions favour different queries, and that the retrieval performance can be significantly enhanced if an appropriate ranking function is selected for each individual query. This thesis proposes Learning to Select (LTS), a novel framework that selectively applies an appropriate ranking function on a per-query basis, regardless of the given query's type and the number of candidate ranking functions. In the learning to select framework, the effectiveness of a ranking function for an unseen query is estimated from the available neighbouring training queries. The proposed framework employs a classification technique (e.g. k-nearest neighbour) to identify neighbouring training queries for an unseen query by using a query feature. In particular, a divergence measure (e.g. Jensen-Shannon), which determines the extent to which a document ranking function alters the scores of an initial ranking of documents for a given query, is proposed for use as a query feature. The ranking function which performs the best on the identified training query set is then chosen for the unseen query. The proposed framework is thoroughly evaluated on two different TREC retrieval tasks (namely, Web search and adhoc search tasks) and on two large standard LETOR feature sets, which contain as many as 64 document features, deriving conclusions concerning the key components of LTS, namely the query feature and the identification of neighbouring queries components. Two different types of experiments are conducted. The first one is to select an appropriate ranking function from a number of candidate ranking functions. The second one is to select multiple appropriate document features from a number of candidate document features, for building a ranking function. Experimental results show that our proposed LTS framework is effective in both selecting an appropriate ranking function and selecting multiple appropriate document features, on a per-query basis. In addition, the retrieval performance is further enhanced when increasing the number of candidates, suggesting the robustness of the learning to select framework. This thesis also demonstrates how the LTS framework can be deployed to other search applications. These applications include the selective integration of a query independent feature into a document weighting scheme (e.g. BM25), the selective estimation of the relative importance of different query aspects in a search diversification task (the goal of the task is to retrieve a ranked list of documents that provides a maximum coverage for a given query, while avoiding excessive redundancy), and the selective application of an appropriate resource for expanding and enriching a given query for document search within an enterprise. The effectiveness of the LTS framework is observed across these search applications, and on different collections, including a large scale Web collection that contains over 50 million documents. This suggests the generality of the proposed learning to select framework. The main contributions of this thesis are the introduction of the LTS framework and the proposed use of divergence measures as query features for identifying similar queries. In addition, this thesis draws insights from a large set of experiments, involving four different standard collections, four different search tasks and large document feature sets. This illustrates the effectiveness, robustness and generality of the LTS framework in tackling various retrieval applications

    Semantic multimedia modelling & interpretation for search & retrieval

    Get PDF
    With the axiomatic revolutionary in the multimedia equip devices, culminated in the proverbial proliferation of the image and video data. Owing to this omnipresence and progression, these data become the part of our daily life. This devastating data production rate accompanies with a predicament of surpassing our potentials for acquiring this data. Perhaps one of the utmost prevailing problems of this digital era is an information plethora. Until now, progressions in image and video retrieval research reached restrained success owed to its interpretation of an image and video in terms of primitive features. Humans generally access multimedia assets in terms of semantic concepts. The retrieval of digital images and videos is impeded by the semantic gap. The semantic gap is the discrepancy between a user’s high-level interpretation of an image and the information that can be extracted from an image’s physical properties. Content- based image and video retrieval systems are explicitly assailable to the semantic gap due to their dependence on low-level visual features for describing image and content. The semantic gap can be narrowed by including high-level features. High-level descriptions of images and videos are more proficient of apprehending the semantic meaning of image and video content. It is generally understood that the problem of image and video retrieval is still far from being solved. This thesis proposes an approach for intelligent multimedia semantic extraction for search and retrieval. This thesis intends to bridge the gap between the visual features and semantics. This thesis proposes a Semantic query Interpreter for the images and the videos. The proposed Semantic Query Interpreter will select the pertinent terms from the user query and analyse it lexically and semantically. The proposed SQI reduces the semantic as well as the vocabulary gap between the users and the machine. This thesis also explored a novel ranking strategy for image search and retrieval. SemRank is the novel system that will incorporate the Semantic Intensity (SI) in exploring the semantic relevancy between the user query and the available data. The novel Semantic Intensity captures the concept dominancy factor of an image. As we are aware of the fact that the image is the combination of various concepts and among the list of concepts some of them are more dominant then the other. The SemRank will rank the retrieved images on the basis of Semantic Intensity. The investigations are made on the LabelMe image and LabelMe video dataset. Experiments show that the proposed approach is successful in bridging the semantic gap. The experiments reveal that our proposed system outperforms the traditional image retrieval systems

    Adaptation des systĂšmes de recherche d'information aux contextes : le cas des requĂȘtes difficiles

    Get PDF
    Le domaine de la recherche d'information (RI) Ă©tudie la façon de trouver des informations pertinentes dans un ou plusieurs corpus, pour rĂ©pondre Ă  un besoin d'information. Dans un SystĂšme de Recherche d'Information (SRI) les informations cherchĂ©es sont des " documents " et un besoin d'information prend la forme d'une " requĂȘte " formulĂ©e par l'utilisateur. La performance d'un SRI est dĂ©pendante de la requĂȘte. Les requĂȘtes pour lesquelles les SRI Ă©chouent (pas ou peu de documents pertinents retrouvĂ©s) sont appelĂ©es dans la littĂ©rature des " requĂȘtes difficiles ". Cette difficultĂ© peut ĂȘtre causĂ©e par l'ambiguĂŻtĂ© des termes, la formulation peu claire de la requĂȘte, le manque de contexte du besoin d'information, la nature et la structure de la collection de documents, etc. Cette thĂšse vise Ă  adapter les systĂšmes de recherche d'information Ă  des contextes, en particulier dans le cadre de requĂȘtes difficiles. Le manuscrit est structurĂ© en cinq chapitres principaux, outre les remerciements, l'introduction gĂ©nĂ©rale et les conclusions et perspectives. Le premier chapitre reprĂ©sente une introduction Ă  la RI. Nous dĂ©veloppons le concept de pertinence, les modĂšles de recherche de la littĂ©rature, l'expansion de requĂȘtes et le cadre d'Ă©valuation utilisĂ© dans les expĂ©rimentations qui ont servi Ă  valider nos propositions. Chacun des chapitres suivants prĂ©sente une de nos contributions. Les chapitres posent les problĂšmes, indiquent l'Ă©tat de l'art, nos propositions thĂ©oriques et leur validation sur des collections de rĂ©fĂ©rence. Dans le chapitre deux, nous prĂ©sentons nos recherche sur la prise en compte du caractĂšre ambigu des requĂȘtes. L'ambiguĂŻtĂ© des termes des requĂȘtes peut en effet conduire Ă  une mauvaise sĂ©lection de documents par les moteurs. Dans l'Ă©tat de l'art, les mĂ©thodes de dĂ©sambiguĂŻsation qui donnent des bonnes performances sont supervisĂ©es, mais ce type de mĂ©thodes n'est pas applicable dans un contexte rĂ©el de RI, car elles nĂ©cessitent de l'information normalement indisponible. De plus, dans la littĂ©rature, la dĂ©sambiguĂŻsation de termes pour la RI est dĂ©clarĂ©e comme sous optimale. Dans ce contexte, nous proposons une mĂ©thode de dĂ©sambiguĂŻsation de requĂȘtes non-supervisĂ©e et montrons son efficacitĂ©. Notre approche est interdisciplinaire, entre les domaines du traitement automatique du langage et la RI. L'objectif de la mĂ©thode de dĂ©sambiguĂŻsation non-supervisĂ©e que nous avons mise au point est de donner plus d'importance aux documents retrouvĂ©s par le moteur de recherche qui contient les mots de la requĂȘte avec les sens identifiĂ©s par la dĂ©sambigĂŒisation. Ce changement d'ordre des documents permet d'offrir une nouvelle liste qui contient plus de documents potentiellement pertinents pour l'utilisateur. Nous avons testĂ© cette mĂ©thode de rĂ©-ordonnancement des documents aprĂšs dĂ©sambigĂŒisation en utilisant deux techniques de classification diffĂ©rentes (NaĂŻve Bayes [Chifu et Ionescu, 2012] et classification spectrale [Chifu et al., 2015]), sur trois collections de documents et des requĂȘtes de la compĂ©tition TREC (TREC7, TREC8, WT10G). Nous avons montrĂ© que la mĂ©thode de dĂ©sambigĂŒisation donne de bons rĂ©sultats dans le cas oĂč peu de documents pertinents sont retrouvĂ©s par le moteur de recherche (7,9% d'amĂ©lioration par rapport aux mĂ©thodes de l'Ă©tat de l'art). Dans le chapitre trois, nous prĂ©sentons le travail focalisĂ© sur la prĂ©diction de la difficultĂ© des requĂȘtes. En effet, si l'ambigĂŒitĂ© est un facteur de difficultĂ©, il n'est pas le seul. Nous avons complĂ©tĂ© la palette des prĂ©dicteurs de difficultĂ© en nous appuyant sur l'Ă©tat de l'art. Les prĂ©dicteurs existants ne sont pas suffisamment efficaces et, en consĂ©quence, nous introduisons des nouvelles mesures de prĂ©diction de la difficultĂ© qui combinent les prĂ©dicteurs. Nous proposons Ă©galement une mĂ©thode robuste pour Ă©valuer les prĂ©dicteurs de difficultĂ© des requĂȘtes. En utilisant les combinaisons des prĂ©dicteurs, sur les collections TREC7 et TREC8, nous obtenons une amĂ©lioration de la qualitĂ© de la prĂ©diction de 7,1% par rapport Ă  l'Ă©tat de l'art [Chifu, 2013]. Dans le quatriĂšme chapitre nous nous intĂ©ressons Ă  l'application des mesures de prĂ©diction. Plus prĂ©cisĂ©ment, nous avons proposĂ© une approche sĂ©lective de RI, c'est-Ă -dire que les prĂ©dicteurs sont utilisĂ©s pour dĂ©cider quel moteur de recherche, parmi plusieurs, rĂ©pondrait mieux pour une requĂȘte. Le modĂšle de dĂ©cision est appris par un SVM (SĂ©parateur Ă  Vaste Marge). Nous avons testĂ© notre modĂšle sur des collections de rĂ©fĂ©rence de TREC (Robust, WT10G, GOV2). Les modĂšles appris ont classĂ© les requĂȘtes de test avec plus de 90% d'exactitude. Par ailleurs, les rĂ©sultats de la recherche ont Ă©tĂ© amĂ©liorĂ©s de plus de 11% en termes de performance, comparĂ© Ă  des mĂ©thodes non sĂ©lectives [Chifu et Mothe, 2014]. Dans le dernier chapitre, nous avons traitĂ© une problĂ©matique importante dans le domaine de la RI : l'expansion des requĂȘtes par l'ajout de termes. Il est trĂšs difficile de prĂ©dire les paramĂštres d'expansion ou d'anticiper si une requĂȘte a besoin d'expansion, ou pas. Nous prĂ©sentons notre contribution pour optimiser le paramĂštre lambda dans le cas de RM3 (un modĂšle pseudo-pertinence d'expansion des requĂȘtes), par requĂȘte. Nous avons testĂ© plusieurs hypothĂšses, Ă  la fois avec et sans information prĂ©alable. Nous recherchons la quantitĂ© minimale d'information nĂ©cessaire pour que l'optimisation du paramĂštre d'expansion soit possible. Les rĂ©sultats obtenus ne sont pas satisfaisants, mĂȘme si nous avons utilisĂ© une vaste plage de mĂ©thodes, comme les SVM, la rĂ©gression, la rĂ©gression logistique et les mesures de similaritĂ©. Par consĂ©quent, ces observations peuvent renforcer la conclusion sur la difficultĂ© de ce problĂšme d'optimisation. Les recherches ont Ă©tĂ© menĂ©es non seulement au cours d'une mobilitĂ© de la recherche de trois mois Ă  l'institut Technion de HaĂŻfa, en IsraĂ«l, en 2013, mais aussi par la suite, en gardant le contact avec l'Ă©quipe de Technion. A HaĂŻfa, nous avons travaillĂ© avec le professeur Oren Kurland et la doctorante Anna Shtok. En conclusion, dans cette thĂšse nous avons proposĂ© de nouvelles mĂ©thodes pour amĂ©liorer les performances des systĂšmes de RI, en s'appuyant sur la difficultĂ© des requĂȘtes. Les rĂ©sultats des mĂ©thodes proposĂ©es dans les chapitres deux, trois et quatre montrent des amĂ©liorations importantes et ouvrent des perspectives pour de futures recherches. L'analyse prĂ©sentĂ©e dans le chapitre cinq confirme la difficultĂ© de la problĂ©matique d'optimisation du paramĂštre concernĂ© et incite Ă  creuser plus sur le paramĂ©trage de l'expansion sĂ©lective des requĂȘtesThe field of information retrieval (IR) studies the mechanisms to find relevant information in one or more document collections, in order to satisfy an information need. For an Information Retrieval System (IRS) the information to find is represented by "documents" and the information need takes the form of a "query" formulated by the user. IRS performance depends on queries. Queries for which the IRS fails (little or no relevant documents retrieved) are called in the literature "difficult queries". This difficulty may be caused by term ambiguity, unclear query formulation, the lack of context for the information need, the nature and structure of the document collection, etc. This thesis aims at adapting IRS to contexts, particularly in the case of difficult queries. The manuscript is organized into five main chapters, besides acknowledgements, general introduction, conclusions and perspectives. The first chapter is an introduction to RI. We develop the concept of relevance, the retrieval models from the literature, the query expansion models and the evaluation framework that was employed to validate our proposals. Each of the following chapters presents one of our contributions. Every chapter raises the research problem, indicates the related work, our theoretical proposals and their validation on benchmark collections. In chapter two, we present our research on treating the ambiguous queries. The query term ambiguity can indeed lead to poor document retrieval of documents by the search engine. In the related work, the disambiguation methods that yield good performance are supervised, however such methods are not applicable in a real IR context, as they require the information which is normally unavailable. Moreover, in the literature, term disambiguation for IR is declared under optimal. In this context, we propose an unsupervised query disambiguation method and show its effectiveness. Our approach is interdisciplinary between the fields of natural language processing and IR. The goal of our unsupervised disambiguation method is to give more importance to the documents retrieved by the search engine that contain the query terms with the specific meaning identified by disambiguation. The document re-ranking provides a new document list that contains potentially relevant documents to the user. We tested this document re-ranking method after disambiguation using two different classification techniques (NaĂŻve Bayes [Chifu and Ionescu, 2012] and spectral clustering [Chifu et al., 2015]), over three document collections and queries from the TREC competition (TREC7, TREC8, WT10G). We have shown that the disambiguation method in IR works well in the case of poorly performing queries (7.9% improvement compared to the methods of the state of the art). In chapter three, we present the work focused on query difficulty prediction. Indeed, if the ambiguity is a difficulty factor, it is not the only one. We completed the range of predictors of difficulty by relying on the state of the art. Existing predictors are not sufficiently effective and therefore we introduce new difficulty prediction measures that combine predictors. We also propose a robust method to evaluate difficulty predictors. Using predictor combinations, on TREC7 and TREC8 collections, we obtain an improvement of 7.1% in terms of prediction quality, compared to the state of the art [Chifu, 2013]. In the fourth chapter we focus on the application of difficulty predictors. Specifically, we proposed a selective IR approach, that is to say, predictors are employed to decide which search engine, among many, would perform better for a query. The decision model is learned by SVM (Support Vector Machine). We tested our model on TREC benchmark collections (Robust, WT10G, GOV2). The learned model classified the test queries with over 90% accuracy. Furthermore, the research results were improved by more than 11% in terms of performance, compared to non-selective methods [Chifu and Mothe, 2014]. In the last chapter, we treated an important issue in the field of IR: the query expansion by adding terms. It is very difficult to predict the expansion parameters or to anticipate whether a query needs the expansion or not. We present our contribution to optimize the lambda parameter in the case of RM3 (a pseudo-relevance model for query expansion), per query. We tested several hypotheses, both with and without prior information. We are searching for the minimum amount of information necessary in order for the optimization of the expansion parameter to be possible. The results are not satisfactory, even though we used a wide range of methods such as SVM, regression, logistic regression and similarity measures. Therefore, these findings may reinforce the conclusion regarding the difficulty of this optimization problem. The research was conducted not only during a mobility research three months at the Technion Institute in Haifa, Israel, in 2013, but thereafter, keeping in touch with the team of Technion. In Haifa, we worked with Professor Oren Kurland and PhD student Anna Shtok. In conclusion, in this thesis we proposed new methods to improve the performance of IRS, based on the query difficulty. The results of the methods proposed in chapters two, three and four show significant improvements and open perspectives for future research. The analysis in chapter five confirms the difficulty of the optimization problem of the concerned parameter and encourages thorough investigation on selective query expansion setting

    Transfer learning for information retrieval

    Get PDF
    The lack of relevance labels is increasingly challenging and presents a bottleneck in the training of reliable learning-to-rank (L2R) models. Obtaining relevance labels using human judgment is expensive and even impossible in some scenarios. Previous research has studied different approaches to solving the problem including generating relevance labels by crowdsourcing and active learning. Recent studies have started to find ways to reuse knowledge from a related collection to help the ranking in a new collection. However, the effectiveness of a ranking function trained in one collection may be degraded when used in another collection due to the generalization issues of machine learning. Transfer learning involves a set of algorithms that are used to train or adapt a model for a target collection without sucient training labels by transferring knowledge from a related source collection with abundant labels. Transfer learning can also be applied to L2R to help train ranking functions for a new task by reusing data from a related collection while minimizing the generalization gap. Some attempts have been made to apply transfer learning techniques on L2R tasks. This thesis investigates different approaches to transfer learning methods for L2R, which are called transfer ranking. However, most of the existing studies on transfer ranking have been focused on the scenario when there are a small but not sucient number of relevance labels. The field of transfer ranking with no target collection labels is still relatively undeveloped. Moreover, the main reason why a transfer ranking solution is needed is that a ranking function trained in the source collection cannot generalize to the target collection, due to the differences in the data distribution of the two collections. However, the effect of the data distribution differences on ranking model generalization has not been examined in detail. The focus of this study is the scenario when there are no relevance labels from the new collection (the target collection), but where a related collection (the target collection) has an abundant amount of training data and labels. In this thesis, we first demonstrate the generalization gap of different L2R algorithms when the distribution of the source and target collections are different in multiple ways, and we then develop alternative solutions to tackling the problem, which includes instance weighting algorithms and self-labeling methods. Instance weighting algorithms estimate weights for each training query in the source collection according to the target query distribution and use the weighted objective function to optimize a ranking function for the target collection. The results on different test collections suggest that instance weighting methods, including existing approaches, are not reliable. The self-labeling methods use other approaches to generate imputed relevance labels for queries in the target collection, which look to transfer the ranking knowledge to the target collection by transferring the label knowledge. The algorithms were tested on various transferring scenarios and showed significant effectiveness and consistency. We thus demonstrate that the performance of self-labeling methods can be further improved with a minimal number of calibration labels from the target collection. The algorithms and knowledge developed in this thesis can help solve generic ranking knowledge transfer problems under different scenarios

    Learning to select a ranking function

    No full text
    Abstract. Learning To Rank (LTR) techniques aim to learn an effective document ranking function by combining several document features. While the function learned may be uniformly applied to all queries, many studies have shown that different ranking functions favour different queries, and the retrieval performance can be significantly enhanced if an appropriate ranking function is selected for each individual query. In this paper, we propose a novel Learning To Select framework that selectively applies an appropriate ranking function on a per-query basis. The approach employs a query feature to identify similar training queries for an unseen query. The ranking function which performs the best on this identified training query set is then chosen for the unseen query. In particular, we propose the use of divergence, which measures the extent that a document ranking function alters the scores of an initial ranking of documents for a given query, as a query feature. We evaluate our method using tasks from the TREC Web and Million Query tracks, in combination with the LETOR 3.0 and LETOR 4.0 feature sets. Our experimental results show that our proposed method is effective and robust for selecting an appropriate ranking function on a per-query basis. In particular, it always outperforms three state-of-the-art LTR techniques, namely Ranking SVM, AdaRank, and the automatic feature selection method.
    corecore