956 research outputs found
Recommended from our members
A user-centred approach to information retrieval
A user model is a fundamental component in user-centred information retrieval systems. It enables personalization of a user's search experience. The development of such a model involves three phases: collecting information about each user, representing such information, and integrating the model into a retrieval application. Progress in this area is typically met with privacy and scalability challenges that hinder the ability to synthesize collective knowledge from each user's search behaviour. In this thesis, I propose a framework that addresses each of these three phases. The proposed framework is based on social role theory from the social science literature and at the centre of this theory is the concept of a social position. A social position is a label for a group of users with similar behavioural patterns. Examples of such positions are traveller, patient, movie fan, and computer scientist. In this thesis, a social position acts as a label for users who are expected to have similar interests. The proposed framework does not require real users' data; rather it uses the web as a resource to model users.
The proposed framework offers a data-driven and modular design for each of the three phases of building a user model. First, I present an approach to identify social positions from natural language sentences. I formulate this task as a binary classification task and develop a method to enumerate candidate social positions. The proposed classifier achieves an accuracy score of 85.8%, which indicates that social positions can be identified with good accuracy. Through an inter-annotator agreement study, I further show a reasonable level of agreement between users when identifying social positions.
Second, I introduce a novel topic modelling-based approach to represent each social position as a multinomial distribution over words. This approach estimates a topic from a document collection for each position. To construct such a collection for a particular position, I propose a seeding algorithm that extracts a set of terms relevant to the social position. Coherence-based evaluation shows that the proposed approach learns significantly more coherent representations when compared with a relevance modelling baseline.
Third, I present a diversification approach based on the proposed framework. Diversification algorithms aim to return a result list for a search query that would potentially satisfy users with diverse information needs. I propose to identify social positions that are relevant to a search query. These positions act as an implicit representation of the many possible interpretations of the search query. Then, relevant positions are provided to a diversification technique that proportionally diversifies results based on each social position's importance. I evaluate my approach using four test collections provided by the diversity task of the Text REtrieval Conference (TREC) web tracks for 2009, 2010, 2011, and 2012. Results demonstrate that my proposed diversification approach is effective and provides statistically significant improvements over various implicit diversification approaches.
Fourth, I introduce a session-based search system under the framework of learning to rank. Such a system aims to improve the retrieval performance for a search query using previous user interactions during the search session. I present a method to match a search session to its most relevant social positions based on the session's interaction data. I then suggest identifying related sessions from query logs that are likely to be issued by users with similar information needs. Novel learning features are then estimated from the session's social positions, related sessions, and interaction data. I evaluate the proposed system using four test collections from the TREC session track. This approach achieves state-of-the-art results compared with effective session-based search systems. I demonstrate that such a strong performance is mainly attributed to features that are derived from social positions' data
Hashing for Multimedia Similarity Modeling and Large-Scale Retrieval
In recent years, the amount of multimedia data such as images, texts, and videos have been growing rapidly on the Internet. Motivated by such trends, this thesis is dedicated to exploiting hashing-based solutions to reveal multimedia data correlations and support intra-media and inter-media similarity search among huge volumes of multimedia data. We start by investigating a hashing-based solution for audio-visual similarity modeling and apply it to the audio-visual sound source localization problem. We show that synchronized signals in audio and visual modalities demonstrate similar temporal changing patterns in certain feature spaces. We propose to use a permutation-based random hashing technique to capture the temporal order dynamics of audio and visual features by hashing them along the temporal axis into a common Hamming space. In this way, the audio-visual correlation problem is transformed into a similarity search problem in the Hamming space. Our hashing-based audio-visual similarity modeling has shown superior performances in the localization and segmentation of sounding objects in videos. The success of the permutation-based hashing method motivates us to generalize and formally define the supervised ranking-based hashing problem, and study its application to large-scale image retrieval. Specifically, we propose an effective supervised learning procedure to learn optimized ranking-based hash functions that can be used for large-scale similarity search. Compared with the randomized version, the optimized ranking-based hash codes are much more compact and discriminative. Moreover, it can be easily extended to kernel space to discover more complex ranking structures that cannot be revealed in linear subspaces. Experiments on large image datasets demonstrate the effectiveness of the proposed method for image retrieval. We further studied the ranking-based hashing method for the cross-media similarity search problem. Specifically, we propose two optimization methods to jointly learn two groups of linear subspaces, one for each media type, so that features\u27 ranking orders in different linear subspaces maximally preserve the cross-media similarities. Additionally, we develop this ranking-based hashing method in the cross-media context into a flexible hashing framework with a more general solution. We have demonstrated through extensive experiments on several real-world datasets that the proposed cross-media hashing method can achieve superior cross-media retrieval performances against several state-of-the-art algorithms. Lastly, to make better use of the supervisory label information, as well as to further improve the efficiency and accuracy of supervised hashing, we propose a novel multimedia discrete hashing framework that optimizes an instance-wise loss objective, as compared to the pairwise losses, using an efficient discrete optimization method. In addition, the proposed method decouples the binary codes learning and hash function learning into two separate stages, thus making the proposed method equally applicable for both single-media and cross-media search. Extensive experiments on both single-media and cross-media retrieval tasks demonstrate the effectiveness of the proposed method
Proceedings of the Workshop on Change of Representation and Problem Reformulation
The proceedings of the third Workshop on Change of representation and Problem Reformulation is presented. In contrast to the first two workshops, this workshop was focused on analytic or knowledge-based approaches, as opposed to statistical or empirical approaches called 'constructive induction'. The organizing committee believes that there is a potential for combining analytic and inductive approaches at a future date. However, it became apparent at the previous two workshops that the communities pursuing these different approaches are currently interested in largely non-overlapping issues. The constructive induction community has been holding its own workshops, principally in conjunction with the machine learning conference. While this workshop is more focused on analytic approaches, the organizing committee has made an effort to include more application domains. We have greatly expanded from the origins in the machine learning community. Participants in this workshop come from the full spectrum of AI application domains including planning, qualitative physics, software engineering, knowledge representation, and machine learning
Proceedings of the XIII Global Optimization Workshop: GOW'16
[Excerpt] Preface: Past Global Optimization Workshop shave been held in Sopron (1985 and 1990), Szeged (WGO, 1995), Florence (GOโ99, 1999), Hanmer Springs (Letโs GO, 2001), Santorini (Frontiers in GO, 2003), San Josรฉ (Goโ05, 2005), Mykonos (AGOโ07, 2007), Skukuza (SAGOโ08, 2008), Toulouse (TOGOโ10, 2010), Natal (NAGOโ12, 2012) and Mรกlaga (MAGOโ14, 2014) with the aim of stimulating discussion between senior and junior researchers on the topic of Global Optimization. In 2016, the XIII Global Optimization Workshop (GOWโ16) takes place in Braga and is organized by three researchers from the University of Minho. Two of them belong to the Systems Engineering and Operational Research Group from the Algoritmi Research Centre and the other to the Statistics, Applied Probability and Operational Research Group from the Centre of Mathematics. The event received more than 50 submissions from 15 countries from Europe, South America and North America. We want to express our gratitude to the invited speaker Panos Pardalos for accepting the invitation and sharing his expertise, helping us to meet the workshop objectives. GOWโ16 would not have been possible without the valuable contribution from the authors and the International Scienti๏ฌc Committee members. We thank you all. This proceedings book intends to present an overview of the topics that will be addressed in the workshop with the goal of contributing to interesting and fruitful discussions between the authors and participants. After the event, high quality papers can be submitted to a special issue of the Journal of Global Optimization dedicated to the workshop. [...
๊ณต๋ ๋์กฐ์ ํ์ต์ ์ด์ฉํ ๋น์ง๋ ๋๋ฉ์ธ ์ ์ ๊ธฐ๋ฒ ์ฐ๊ตฌ
ํ์๋
ผ๋ฌธ (์์ฌ) -- ์์ธ๋ํ๊ต ๋ํ์ : ๊ณต๊ณผ๋ํ ์ ๊ธฐยท์ ๋ณด๊ณตํ๋ถ, 2021. 2. ์ค์ฑ๋ก.Domain adaptation is introduced to exploit the label information of source domain when labels are not available for target domain. Previous methods minimized domain discrepancy in a latent space to enable transfer learning. These studies are based on the theoretical analysis that the target error is upper bounded by the sum of source error, the domain discrepancy, and the joint error of the ideal hypothesis. However, feature discriminability is sacrificed while enhancing the feature transferability by matching marginal distributions. In particular, the ideal joint hypothesis error in the target error upper bound, which was previously considered to be minute, has been found to be significant, impairing its theoretical guarantee.
In this paper, to manage the joint error, we propose an alternative upper bound on the target error that explicitly considers it. Based on the theoretical analysis, we suggest a joint optimization framework that combines the source and target domains. To minimize the joint error, we further introduce Joint Contrastive Learning (JCL) that finds class-level discriminative features. With a solid theoretical framework, JCL employs contrastive loss to maximize the mutual information between a feature and its label, which is equivalent to maximizing the Jensen-Shannon divergence between conditional distributions. Extensive experiments on domain adaptation datasets demonstrate that JCL outperforms existing state-of-the-art methods.๋๋ฉ์ธ ์ ์ ๊ธฐ๋ฒ์ ํ๊ฒ ๋๋ฉ์ธ์ ๋ผ๋ฒจ ์ ๋ณด๊ฐ ์๋ ์ํฉ์์ ๋น์ทํ ๋๋ฉ์ธ์ธ ์์ค ๋๋ฉ์ธ์ ๋ผ๋ฒจ ์ ๋ณด๋ฅผ ํ์ฉํ๊ธฐ ์ํด ๊ฐ๋ฐ๋์๋ค. ๊ธฐ์กด์ ๋ฐฉ๋ฒ๋ก ๋ค์ ์ ์ฌ ๊ณต๊ฐ์์ ๋๋ฉ์ธ๋ค ์ฌ์ด์ ๋ถํฌ ์ฐจ์ด๋ฅผ ์ค์์ผ๋ก์จ ์ ์ด ํ์ต์ด ๊ฐ๋ฅํ๊ฒ ํ์๋ค. ์ด๋ฌํ ๊ธฐ๋ฒ๋ค์ ์์ค ๋๋ฉ์ธ์ ์๋ฌ์จ, ๋๋ฉ์ธ ๊ฐ ๋ถํฌ ์ฐจ์ด, ๊ทธ๋ฆฌ๊ณ ์ ๋๋ฉ์ธ์์ ์ด์์ ์ธ ๋ถ๋ฅ๊ธฐ์ ์๋ฌ์จ์ ํฉ์ด ํ๊ฒ ๋๋ฉ์ธ์ ์๋ฌ์จ์ ์๊ณ๊ฐ ๋๋ค๋ ์ด๋ก ์ ๋ฐํ์ผ๋ก ํ๋ค. ๊ทธ๋ฌ๋ ๋๋ฉ์ธ๋ค ์ฌ์ด์ ๋ถํฌ ์ฐจ์ด๋ฅผ ์ค์ด๋ ๋ฐฉ๋ฒ๋ค์ ๋์์ ์ ์ฌ ๊ณต๊ฐ์์ ์๋ก ๋ค๋ฅธ ๋ผ๋ฒจ์ ๊ฐ๋ ๋ฐ์ดํฐ๋ค ์ฌ์ด์ ๊ตฌ๋ณ์ฑ์ ๊ฐ์์์ผฐ๋ค. ํนํ, ์์ ๊ฒ์ด๋ผ ์๊ฐ๋๋ ์ ๋๋ฉ์ธ์์ ์ด์์ ์ธ ๋ถ๋ฅ๊ธฐ์ ์๋ฌ์จ์ด ํฐ ๊ฒ์ผ๋ก ๋ํ๋ฌ๋ค.
๋ณธ ๋
ผ๋ฌธ์์๋ ๊ธฐ์กด์ ์ด๋ก ์์๋ ๋ค๋ฃจ์ง ์์ ์ ๋๋ฉ์ธ์์ ๋ถ๋ฅ๊ธฐ์ ์๋ฌ์จ์ ์กฐ์ ํ ์ ์๊ฒํ๊ธฐ ์ํด ์๋ก์ด ์ด๋ก ์ ์ ์ํ๋ค. ์ด ์ด๋ก ์ ๋ฐฐ๊ฒฝ์ ๋ฐํ์ผ๋ก ์์ค ๋๋ฉ์ธ๊ณผ ํ๊ฒ ๋๋ฉ์ธ์ ํจ๊ป ํ์ตํ๋ ๊ณต๋ ๋์กฐ์ ๋ฐฉ๋ฒ์ ์๊ฐํ๋ค. ๋ณธ ๊ณต๋ ๋์กฐ์ ํ์ต ๋ฐฉ๋ฒ์์๋ ๊ฐ ๋ผ๋ฒจ๋ณ๋ก ๊ตฌ๋ถ๋๋ ์ ์ฌ ๊ณต๊ฐ์ ํ์ตํ๊ธฐ ์ํด ๊ฐ ๋ฐ์ดํฐ์ ํน์ง๊ณผ ๋ผ๋ฒจ ์ฌ์ด์ ์ํธ ์ ๋ณด๋์ ์ต๋ํํ๋ค. ์ด ๊ฐ ๋ฐ์ดํฐ์ ํน์ง๊ณผ ๋ผ๋ฒจ ์ฌ์ด์ ์ํธ ์ ๋ณด๋์ ๊ฐ ๋ผ๋ฒจ ๋ถํฌ ์ฌ์ด์ ์ ์ผ-์ค๋
ผ ๊ฑฐ๋ฆฌ์ ๊ฐ์ผ๋ฏ๋ก ์ด๋ฅผ ์ต๋ํํ๋ ๊ฒ์ ๊ณง ๋ผ๋ฒจ๋ค์ด ์ ๊ตฌ๋ณ๋๋ ์ ์ฌ ๊ณต๊ฐ์ ํ์ตํ๋ ๊ฒ์ด๋ค. ๋ง์ง๋ง์ผ๋ก ๊ณต๋ ๋์กฐ์ ํ์ต ๋ฐฉ๋ฒ์ ์ฌ๋ฌ ๋ฐ์ดํฐ ์
์ ์ ์ฉํ์ฌ ๊ธฐ์กด ๋ฐฉ๋ฒ๋ก ๋ค๊ณผ ๋น๊ตํ์๋ค.1 Introduction 1
2 Background 4
2.1 Domain Adaptation . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.1.1 Problem Setting and Notations . . . . . . . . . . . . . . . . . 4
2.1.2 Theoretical Background . . . . . . . . . . . . . . . . . . . . 5
2.2 Approaches for Domain Adaptation . . . . . . . . . . . . . . . . . . 6
2.2.1 Marginal Distribution Alignment Based Approaches . . . . . 6
2.2.2 Conditional Distribution Matching Approaches . . . . . . . . 7
2.3 Contrastive Learning . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3 Method 10
3.1 An Alternative Upper Bound . . . . . . . . . . . . . . . . . . . . . . 10
3.2 Joint Contrastive Learning . . . . . . . . . . . . . . . . . . . . . . . 14
3.2.1 Theoretical Guarantees . . . . . . . . . . . . . . . . . . . . . 14
3.2.2 Generalization to Multiclass Classification . . . . . . . . . . 17
3.2.3 Training Procedure . . . . . . . . . . . . . . . . . . . . . . . 19
4 Experiments 24
4.1 Datasets and Baselines . . . . . . . . . . . . . . . . . . . . . . . . . 24
4.2 Implementation Details . . . . . . . . . . . . . . . . . . . . . . . . . 26
4.3 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
4.4 Ablation Studies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
5 Conclusion 35
Abstract (In Korean) 45Maste
Reformulation and Decomposition: Multitask learning approaches to Long Document Problems
Recent advances in Natural Language Processing (NLP) have led to success across a wide range of tasks including machine translation, summarization, and classification. Yet, the field still faces major challenges. This thesis addresses two key under-researched areas: the absence of general multitask learning capabilities, and the inability to scale to long, complex documents. Firstly, this thesis explores a form of multitasking where NLP tasks are reformulated as question answering problems. I examine existing models and measure their robustness to paraphrasing of their input. I contribute an annotated dataset which enables detailed analysis of model failures as well as evaluating methods for improving model robustness. Secondly, a set of long document tasks; MuLD, is introduced which forms a benchmark for evaluating the performance of models on large inputs with long-range dependencies. I show that this is a challenging task for baseline models. I then design an approach using task-decomposition to provide an interpretable solution which easily allows for multitask learning. I then explore how these themes of task reformulation for multitask learning, and task-decomposition for long inputs can be applied to other modalities. I show how visual modelling: a visual analogue of language modelling, can be used to predict missing frames from videos of simple physics simulations, and probe what knowledge about the physical world this induces in such models. Finally, I demonstrate how this task can be used to unite vision and NLP using the same framework, describing how task-reformulation and task-decomposition can be used for this purpose
Semantic multimedia modelling & interpretation for search & retrieval
With the axiomatic revolutionary in the multimedia equip devices, culminated in the proverbial proliferation of the image and video data. Owing to this omnipresence and progression, these data become the part of our daily life. This devastating data production rate accompanies with a predicament of surpassing our potentials for acquiring this data. Perhaps one of the utmost prevailing problems of this digital era is an information plethora.
Until now, progressions in image and video retrieval research reached restrained success owed to its interpretation of an image and video in terms of primitive features. Humans generally access multimedia assets in terms of semantic concepts. The retrieval of digital images and videos is impeded by the semantic gap. The semantic gap is the discrepancy between a userโs high-level interpretation of an image and the information that can be extracted from an imageโs physical properties. Content- based image and video retrieval systems are explicitly assailable to the semantic gap due to their dependence on low-level visual features for describing image and content. The semantic gap can be narrowed by including high-level features. High-level descriptions of images and videos are more proficient of apprehending the semantic meaning of image and video content.
It is generally understood that the problem of image and video retrieval is still far from being solved. This thesis proposes an approach for intelligent multimedia semantic extraction for search and retrieval. This thesis intends to bridge the gap between the visual features and semantics. This thesis proposes a Semantic query Interpreter for the images and the videos. The proposed Semantic Query Interpreter will select the pertinent terms from the user query and analyse it lexically and semantically. The proposed SQI reduces the semantic as well as the vocabulary gap between the users and the machine. This thesis also explored a novel ranking strategy for image search and retrieval. SemRank is the novel system that will incorporate the Semantic Intensity (SI) in exploring the semantic relevancy between the user query and the available data. The novel Semantic Intensity captures the concept dominancy factor of an image. As we are aware of the fact that the image is the combination of various concepts and among the list of concepts some of them are more dominant then the other. The SemRank will rank the retrieved images on the basis of Semantic Intensity.
The investigations are made on the LabelMe image and LabelMe video dataset. Experiments show that the proposed approach is successful in bridging the semantic gap. The experiments reveal that our proposed system outperforms the traditional image retrieval systems
Knowledge-directed intelligent information retrieval for research funding.
Thesis (M.Sc.)- University of Natal, Pietermaritzburg, 2001.Researchers have always found difficulty in attaining funding from the National Research Foundation (NRF) for new research interests. The field of Artificial Intelligence (AI) holds the promise of improving the matching of research proposals to funding sources in the area of Intelligent Information Retrieval (IIR). IIR is a fairly new AI technique that has evolved from the traditional IR systems to solve real-world problems. Typically, an IIR system contains three main components, namely, a knowledge base, an inference engine and a user-interface. Due to its inferential capabilities. IIR has been found to be applicable to domains for which traditional techniques, such as the use of databases, have not been well suited. This applicability has led it to become a viable AI technique from both, a research and an application perspective. This dissertation concentrates on researching and implementing an IIR system in LPA Prolog, that we call FUND, to assist in the matching of research proposals of prospective researchers to funding sources within the National Research Foundation (NRF). FUND'S reasoning strategy for its inference engine is backward chaining that carries out a depth-first search over its knowledge representation structure, namely, a semantic network. The distance constraint of the Constrained Spreading Activation (CSA) technique is incorporated within the search strategy to help prune non-relevant returns by FUND. The evolution of IIR from IR was covered in detail. Various reasoning strategies and knowledge representation schemes were reviewed to find the combination that best suited the problem domain and programming language chosen. FUND accommodated a depth 4, depth 5 and an exhaustive search algorithm. FUND'S effectiveness was tested, in relation to the different searches with respect to their precision and recall ability and in comparison to other similar systems. FUND'S performance in providing researchers with better funding advice in the South African situation proved to be favourably comparable to other similar systems elsewhere
- โฆ