8 research outputs found

    Dynamic match kernel with deep convolutional features for image retrieval

    Get PDF
    For image retrieval methods based on bag of visual words, much attention has been paid to enhancing the discriminative powers of the local features. Although retrieved images are usually similar to a query in minutiae, they may be significantly different from a semantic perspective, which can be effectively distinguished by convolutional neural networks (CNN). Such images should not be considered as relevant pairs. To tackle this problem, we propose to construct a dynamic match kernel by adaptively calculating the matching thresholds between query and candidate images based on the pairwise distance among deep CNN features. In contrast to the typical static match kernel which is independent to the global appearance of retrieved images, the dynamic one leverages the semantical similarity as a constraint for determining the matches. Accordingly, we propose a semantic-constrained retrieval framework by incorporating the dynamic match kernel, which focuses on matched patches between relevant images and filters out the ones for irrelevant pairs. Furthermore, we demonstrate that the proposed kernel complements recent methods, such as hamming embedding, multiple assignment, local descriptors aggregation, and graph-based re-ranking, while it outperforms the static one under various settings on off-the-shelf evaluation metrics. We also propose to evaluate the matched patches both quantitatively and qualitatively. Extensive experiments on five benchmark data sets and large-scale distractors validate the merits of the proposed method against the state-of-the-art methods for image retrieval

    Deep Image Retrieval: A Survey

    Get PDF
    In recent years a vast amount of visual content has been generated and shared from various fields, such as social media platforms, medical images, and robotics. This abundance of content creation and sharing has introduced new challenges. In particular, searching databases for similar content, i.e.content based image retrieval (CBIR), is a long-established research area, and more efficient and accurate methods are needed for real time retrieval. Artificial intelligence has made progress in CBIR and has significantly facilitated the process of intelligent search. In this survey we organize and review recent CBIR works that are developed based on deep learning algorithms and techniques, including insights and techniques from recent papers. We identify and present the commonly-used benchmarks and evaluation methods used in the field. We collect common challenges and propose promising future directions. More specifically, we focus on image retrieval with deep learning and organize the state of the art methods according to the types of deep network structure, deep features, feature enhancement methods, and network fine-tuning strategies. Our survey considers a wide variety of recent methods, aiming to promote a global view of the field of instance-based CBIR.Comment: 20 pages, 11 figure

    Computer vision beyond the visible : image understanding through language

    Get PDF
    In the past decade, deep neural networks have revolutionized computer vision. High performing deep neural architectures trained for visual recognition tasks have pushed the field towards methods relying on learned image representations instead of hand-crafted ones, in the seek of designing end-to-end learning methods to solve challenging tasks, ranging from long-lasting ones such as image classification to newly emerging tasks like image captioning. As this thesis is framed in the context of the rapid evolution of computer vision, we present contributions that are aligned with three major changes in paradigm that the field has recently experienced, namely 1) the power of re-utilizing deep features from pre-trained neural networks for different tasks, 2) the advantage of formulating problems with end-to-end solutions given enough training data, and 3) the growing interest of describing visual data with natural language rather than pre-defined categorical label spaces, which can in turn enable visual understanding beyond scene recognition. The first part of the thesis is dedicated to the problem of visual instance search, where we particularly focus on obtaining meaningful and discriminative image representations which allow efficient and effective retrieval of similar images given a visual query. Contributions in this part of the thesis involve the construction of sparse Bag-of-Words image representations from convolutional features from a pre-trained image classification neural network, and an analysis of the advantages of fine-tuning a pre-trained object detection network using query images as training data. The second part of the thesis presents contributions to the problem of image-to-set prediction, understood as the task of predicting a variable-sized collection of unordered elements for an input image. We conduct a thorough analysis of current methods for multi-label image classification, which are able to solve the task in an end-to-end manner by simultaneously estimating both the label distribution and the set cardinality. Further, we extend the analysis of set prediction methods to semantic instance segmentation, and present an end-to-end recurrent model that is able to predict sets of objects (binary masks and categorical labels) in a sequential manner. Finally, the third part of the dissertation takes insights learned in the previous two parts in order to present deep learning solutions to connect images with natural language in the context of cooking recipes and food images. First, we propose a retrieval-based solution in which the written recipe and the image are encoded into compact representations that allow the retrieval of one given the other. Second, as an alternative to the retrieval approach, we propose a generative model to predict recipes directly from food images, which first predicts ingredients as sets and subsequently generates the rest of the recipe one word at a time by conditioning both on the image and the predicted ingredients.En l'煤ltima d猫cada, les xarxes neuronals profundes han revolucionat el camp de la visi贸 per computador. Els resultats favorables obtinguts amb arquitectures neuronals profundes entrenades per resoldre tasques de reconeixement visual han causat un canvi de paradigma cap al disseny de m猫todes basats en representacions d'imatges apreses de manera autom脿tica, deixant enrere les t猫cniques tradicionals basades en l'enginyeria de representacions. Aquest canvi ha perm猫s l'aparici贸 de t猫cniques basades en l'aprenentatge d'extrem a extrem (end-to-end), capaces de resoldre de manera efectiva molts dels problemes tradicionals de la visi贸 per computador (e.g. classificaci贸 d'imatges o detecci贸 d'objectes), aix铆 com nous problemes emergents com la descripci贸 textual d'imatges (image captioning). Donat el context de la r脿pida evoluci贸 de la visi贸 per computador en el qual aquesta tesi s'emmarca, presentem contribucions alineades amb tres dels canvis m茅s importants que la visi贸 per computador ha experimentat recentment: 1) la reutilitzaci贸 de representacions extretes de models neuronals pre-entrenades per a tasques auxiliars, 2) els avantatges de formular els problemes amb solucions end-to-end entrenades amb grans bases de dades, i 3) el creixent inter猫s en utilitzar llenguatge natural en lloc de conjunts d'etiquetes categ貌riques pre-definits per descriure el contingut visual de les imatges, facilitant aix铆 l'extracci贸 d'informaci贸 visual m茅s enll脿 del reconeixement de l'escena i els elements que la composen La primera part de la tesi est脿 dedicada al problema de la cerca d'imatges (image retrieval), centrada especialment en l'obtenci贸 de representacions visuals significatives i discriminat貌ries que permetin la recuperaci贸 eficient i efectiva d'imatges donada una consulta formulada amb una imatge d'exemple. Les contribucions en aquesta part de la tesi inclouen la construcci贸 de representacions Bag-of-Words a partir de descriptors locals obtinguts d'una xarxa neuronal entrenada per classificaci贸, aix铆 com un estudi dels avantatges d'utilitzar xarxes neuronals per a detecci贸 d'objectes entrenades utilitzant les imatges d'exemple, amb l'objectiu de millorar les capacitats discriminat貌ries de les representacions obtingudes. La segona part de la tesi presenta contribucions al problema de predicci贸 de conjunts a partir d'imatges (image to set prediction), ent猫s com la tasca de predir una col路lecci贸 no ordenada d'elements de longitud variable donada una imatge d'entrada. En aquest context, presentem una an脿lisi exhaustiva dels m猫todes actuals per a la classificaci贸 multi-etiqueta d'imatges, que s贸n capa莽os de resoldre la tasca de manera integral calculant simult脿niament la distribuci贸 probabil铆stica sobre etiquetes i la cardinalitat del conjunt. Seguidament, estenem l'an脿lisi dels m猫todes de predicci贸 de conjunts a la segmentaci贸 d'inst脿ncies sem脿ntiques, presentant un model recurrent capa莽 de predir conjunts d'objectes (representats per m脿scares bin脿ries i etiquetes categ貌riques) de manera seq眉encial. Finalment, la tercera part de la tesi est茅n els coneixements apresos en les dues parts anteriors per presentar solucions d'aprenentatge profund per connectar imatges amb llenguatge natural en el context de receptes de cuina i imatges de plats cuinats. En primer lloc, proposem una soluci贸 basada en algoritmes de cerca, on la recepta escrita i la imatge es codifiquen amb representacions compactes que permeten la recuperaci贸 d'una donada l'altra. En segon lloc, com a alternativa a la soluci贸 basada en algoritmes de cerca, proposem un model generatiu capa莽 de predir receptes (compostes pels seus ingredients, predits com a conjunts, i instruccions) directament a partir d'imatges de menjar.Postprint (published version

    Large scale visual search

    Get PDF
    With the ever-growing amount of image data on the web, much attention has been devoted to large scale image search. It is one of the most challenging problems in computer vision for several reasons. First, it must address various appearance transformations such as changes in perspective, rotation and scale existing in the huge amount of image data. Second, it needs to minimize memory requirements and computational cost when generating image representations. Finally, it needs to construct an efficient index space and a suitable similarity measure to reduce the response time to the users. This thesis aims to provide robust image representations that are less sensitive to above mentioned appearance transformations and are suitable for large scale image retrieval. Although this thesis makes a substantial number of contributions to large scale image retrieval, we also presented additional challenges and future research based on the contributions in this thesis.China Scholarship Council (CSC)Computer Systems, Imagery and Medi
    corecore