4 research outputs found

    Hybrid image representation methods for automatic image annotation: a survey

    Get PDF
    In most automatic image annotation systems, images are represented with low level features using either global methods or local methods. In global methods, the entire image is used as a unit. Local methods divide images into blocks where fixed-size sub-image blocks are adopted as sub-units; or into regions by using segmented regions as sub-units in images. In contrast to typical automatic image annotation methods that use either global or local features exclusively, several recent methods have considered incorporating the two kinds of information, and believe that the combination of the two levels of features is beneficial in annotating images. In this paper, we provide a survey on automatic image annotation techniques according to one aspect: feature extraction, and, in order to complement existing surveys in literature, we focus on the emerging image annotation methods: hybrid methods that combine both global and local features for image representation

    Hierarchical Ensemble of Global and Local Classifiers for Face Recognition

    Full text link

    Palmprint Authentication System Based on Local and Global Feature Fusion Using DOST

    Get PDF
    Palmprint is the region between wrist and fingers. In this paper, a palmprint personal identification system is proposed based on the local and global information fusion. The local and global information is critical for the image observation based on the results of the relationship between physical stimuli and perceptions. The local features of the enhanced palmprint are extracted using discrete orthonormal Stockwell transform. The global feature is obtained by reducing the scale of discrete orthonormal Stockwell transform to infinity. The local and global matching distances of the two palmprint images are fused to get the final matching distance of the proposed scheme. The results show that the fusion of local and global features outperforms the existing works on the available three datasets

    Annotation d'images via leur contexte spatio-temporel et les métadonnées du Web

    Get PDF
    En Recherche d'Information (RI), les documents sont classiquement indexĂ©s en fonction de leur contenu, qu'il soit textuel ou multimĂ©dia. Les moteurs de recherche s'appuyant sur ces index sont aujourd'hui des outils performants, rĂ©pandus et indispensables. Ils visent Ă  fournir des rĂ©ponses pertinentes selon le besoin de l'utilisateur, sous forme de textes, images, sons, vidĂ©os, etc. Nos travaux de thĂšse s'inscrivent dans le contexte des documents de type image. Plus prĂ©cisĂ©ment, nous nous sommes intĂ©ressĂ©s aux systĂšmes d'annotation automatique d'images qui permettent d'associer automatiquement des mots-clĂ©s Ă  des images afin de pouvoir ensuite les rechercher par requĂȘte textuelle. Ce type d'annotation cherche Ă  combler les lacunes des approches d'annotation manuelle et semi-automatique. Celles-ci ne sont plus envisageables dans le contexte actuel qui permet Ă  chacun de prendre de nombreuses photos Ă  faible coĂ»t (en lien avec la dĂ©mocratisation des appareils photo numĂ©riques et l'intĂ©gration de capteurs numĂ©riques dans les tĂ©lĂ©phones mobiles). Parmi les diffĂ©rents types de collections d'images existantes (par exemple, mĂ©dicales, satellitaires) dans le cadre de cette thĂšse nous nous sommes intĂ©ressĂ©s aux collections d'images de type paysage (c.-Ă -d. des images qui illustrent des points d'intĂ©rĂȘt touristiques) pour lesquelles nous avons identifiĂ© des dĂ©fis, tels que l'identification des nouveaux descripteurs pour les dĂ©crire et de nouveaux modĂšles pour fusionner ces derniers, l'identification des sources d'information pertinentes et le passage Ă  l'Ă©chelle. Nos contributions portent sur trois principaux volets. En premier lieu, nous nous sommes attachĂ©s Ă  exploiter diffĂ©rents descripteurs qui peuvent influencer la description des images de type paysage : le descripteur de spatialisation (caractĂ©risĂ© par la latitude et la longitude des images), le descripteur de temporalitĂ© (caractĂ©risĂ© par la date et l'heure de la prise de vue) et le descripteur de thĂ©matique (caractĂ©risĂ© par les tags issus des plate formes de partage d'images). Ensuite, nous avons proposĂ© des approches pour modĂ©liser ces descripteurs au regard de statistiques de tags liĂ©es Ă  leur frĂ©quence et raretĂ© et sur des similaritĂ©s spatiale et temporelle. DeuxiĂšmement, nous avons proposĂ© un nouveau processus d'annotation d'images qui vise Ă  identifier les mots-clĂ©s qui dĂ©crivent le mieux les images-requĂȘtes donnĂ©es en entrĂ©e d'un systĂšme d'annotation par un utilisateur. Pour ce faire, pour chaque image-requĂȘte nous avons mis en Ɠuvre des filtres spatial, temporel et spatio-temporel afin d'identifier les images similaires ainsi que leurs tags associĂ©s. Ensuite, nous avons fĂ©dĂ©rĂ© les diffĂ©rents descripteurs dans un modĂšle probabiliste afin de dĂ©terminer les termes qui dĂ©crivent le mieux chaque image-requĂȘte. Enfin, le fait que les contributions prĂ©sentĂ©es ci-dessus s'appuient uniquement sur des informations issues des plateformes de partage d'images (c.-Ă -d. des informations subjectives) a suscitĂ© la question suivante : les informations issues du Web peuvent-elles fournir des termes objectifs pour enrichir les descriptions initiales des images. À cet effet, nous avons proposĂ© une approche basĂ©e sur les techniques d'expansion de requĂȘtes du domaine de la RI. Elle porte essentiellement sur l'Ă©tude de l'impact des diffĂ©rents algorithmes d'expansion, ainsi que sur l'agrĂ©gation des rĂ©sultats fournis par le meilleur algorithme et les rĂ©sultats fournis par le processus d'annotation d'images. Vu qu'il n'existe pas de cadre d'Ă©valuation standard d'annotation automatique d'images, plus particuliĂšrement adaptĂ© aux collections d'images de type paysage, nous avons proposĂ© des cadres d'Ă©valuation appropriĂ©s afin de valider nos contributions. En particulier, les diffĂ©rentes approches proposĂ©es sont Ă©valuĂ©es au regard de la modĂ©lisation des descripteur de spatialisation, de temporalitĂ© et de thĂ©matique. De plus, nous avons validĂ© le processus d'annotation d'images, et nous avons montrĂ© qu'il surpasse en qualitĂ© deux approches d'annotation d'images de la littĂ©rature. Nous avons comparĂ© Ă©galement l'approche d'enrichissement avec le processus d'annotation d'image pour souligner son efficacitĂ© et l'apport des informations issues du Web. Ces expĂ©rimentations ont nĂ©cessitĂ© le prototypage du logiciel AnnoTaGT, qui offre aux utilisateurs un cadre technique pour l'annotation automatique d'images.The documents processed by Information Retrieval (IR) systems are typically indexed according to their contents: Text or multimedia. Search engines based on these indexes aim to provide relevant answers to users' needs in the form of texts, images, sounds, videos, and so on. Our work is related to "image" documents. We are specifically interested in automatic image annotation systems that automatically associate keywords to images. Keywords are subsequently used for search purposes via textual queries. The automatic image annotation task intends to overcome the issues of manual and semi-automatic annotation tasks, as they are no longer feasible in nowadays' context (i.e., the development of digital technologies and the advent of devices, such as smartphones, allowing anyone to take images with a minimal cost). Among the different types of existing image collections (e.g., medical, satellite) in our work we are interested in landscape image collections for which we identified the following challenges: What are the most discriminant features for this type of images ? How to model and how to merge these features ? What are the sources of information that should be considered ? How to manage scalability issues ? The proposed contribution is threefold. First, we use different factors that influence the description of landscape images: The spatial factor (i.e., latitude and longitude of images), the temporal factor (i.e., the time when the images were taken), and the thematic factor (i.e., tags crowdsourced and contributed to image sharing platforms). We propose various techniques to model these factors based on tag frequency, as well as spatial and temporal similarities. The choice of these factors is based on the following assumptions: A tag is all the more relevant for a query-image as it is associated with images located in its close geographical area ; A tag is all the more relevant for a query-image as it is associated with images captured close in time to it ; sourcing concept). Second, we introduce a new image annotation process that recommends the terms that best describe a given query-image provided by a user. For each query-image we rely on spatial, temporal, and spatio-temporal filters to identify similar images along with their tags. Then, the different factors are merged through a probabilistic model to boost the terms best describing each query-image. Third, the contributions presented above are only based on information extracted from image photo sharing platforms (i.e., subjective information). This raised the following research question: Can the information extracted from the Web provide objective terms useful to enrich the initial description of images? We tackle this question by introducing an approach relying on query expansion techniques developed in IR. As there is no standard evaluation protocol for the automatic image annotation task tailored to landscape images, we designed various evaluation protocols to validate our contributions. We first evaluated the approaches defined to model the spatial, temporal, and thematic factors. Then, we validated the annotation image process and we showed that it yields significant improvement over two state-of-the-art baselines. Finally, we assessed the effectiveness of tag expansion through Web sources and showed its contribution to the image annotation process. These experiments are complemented by the image annotation prototype AnnoTaGT, which provides users with an operational framework for automatic image annotation
    corecore