13 research outputs found

    Face Retrieval Using Image Moments and Genetic Algorithm

    Get PDF
    Content-based image retrieval has been developed in numerous fields. This provides more active management and retrieval of images than the keyword-based method. So the content based image retrieval becomes one of the liveliest researches in the past few years.In a given set of objects, the retrieval of information suggests solutions to search forthose in response to a particular description. The set of objects which can be considered are documents,images, videos, or sounds. Moments can be viewed as powerful image descriptors that capture global characteristics of an image. The magnitude of the moment coefficients is said to be invariant under geometrical transformations like rotation which makes them suitable for most of the recognition applications. This paper presents a method to retrieve a multi-view face from a large face database according to face image moments and genetic algorithm. The GA is preferred for its power and because it can be usedwithout any specificinformation of the domain. The experimental results concludes thatusing GA gives a good performance and it decreases the average search time to (56.44milliseconds) compared with (891.6 milliseconds) for traditional search

    A unified learning framework for content based medical image retrieval using a statistical model

    Get PDF
    AbstractThis paper presents a unified learning framework for heterogeneous medical image retrieval based on a Full Range Autoregressive Model (FRAR) with the Bayesian approach (BA). Using the unified framework, the color autocorrelogram, edge orientation autocorrelogram (EOAC) and micro-texture information of medical images are extracted. The EOAC is constructed in HSV color space, to circumvent the loss of edges due to spectral and chromatic variations. The proposed system employed adaptive binary tree based support vector machine (ABTSVM) for efficient and fast classification of medical images in feature vector space. The Manhattan distance measure of order one is used in the proposed system to perform a similarity measure in the classified and indexed feature vector space. The precision and recall (PR) method is used as a measure of performance in the proposed system. Short-term based relevance feedback (RF) mechanism is also adopted to reduce the semantic gap. The Experimental results reveal that the retrieval performance of the proposed system for heterogeneous medical image database is better than the existing systems at low computational and storage cost

    Texture Based Image retrieval using Human interactive Genetic Algorithm

    Get PDF
    Content-based image retrieval has been keenly calculated in numerous fields. This provides more active management and retrieval of images than the keyword-based method. So the content based image retrieval has become one of the liveliest researches in the past few years. As earlier, we were using the text-based approach where it initiate very boring and hard task for solving the purpose of image retrieval. But the CBIR is the method where there are several methodologies are available and the task of image retrieval becomes well easier. In this, there are specific effective methods for CBIR are discussed and the relative study is made. However most of the proposed methods emphasize on finding the best representation for diverse image features. Here, the user-oriented mechanism for CBIR method based on an interactivegenetic algorithm (IGA) is proposed. Color attributes likethe mean value, the standard deviation, and the image bitmap of a color image are used as the features for retrieval. In addition, the entropy based on the gray level co-occurrence matrix and the edge histograms of an image are too considered as the texture features

    Multifeature analysis and semantic context learning for image classification

    Get PDF
    This article introduces an image classification approach in which the semantic context of images and multiple low-level visual features are jointly exploited. The context consists of a set of semantic terms defining the classes to be associated to unclassified images. Initially, a multiobjective optimization technique is used to define a multifeature fusion model for each semantic class. Then, a Bayesian learning procedure is applied to derive a context model representing relationships among semantic classes. Finally, this ..

    Texture Structure Analysis

    Get PDF
    abstract: Texture analysis plays an important role in applications like automated pattern inspection, image and video compression, content-based image retrieval, remote-sensing, medical imaging and document processing, to name a few. Texture Structure Analysis is the process of studying the structure present in the textures. This structure can be expressed in terms of perceived regularity. Our human visual system (HVS) uses the perceived regularity as one of the important pre-attentive cues in low-level image understanding. Similar to the HVS, image processing and computer vision systems can make fast and efficient decisions if they can quantify this regularity automatically. In this work, the problem of quantifying the degree of perceived regularity when looking at an arbitrary texture is introduced and addressed. One key contribution of this work is in proposing an objective no-reference perceptual texture regularity metric based on visual saliency. Other key contributions include an adaptive texture synthesis method based on texture regularity, and a low-complexity reduced-reference visual quality metric for assessing the quality of synthesized textures. In order to use the best performing visual attention model on textures, the performance of the most popular visual attention models to predict the visual saliency on textures is evaluated. Since there is no publicly available database with ground-truth saliency maps on images with exclusive texture content, a new eye-tracking database is systematically built. Using the Visual Saliency Map (VSM) generated by the best visual attention model, the proposed texture regularity metric is computed. The proposed metric is based on the observation that VSM characteristics differ between textures of differing regularity. The proposed texture regularity metric is based on two texture regularity scores, namely a textural similarity score and a spatial distribution score. In order to evaluate the performance of the proposed regularity metric, a texture regularity database called RegTEX, is built as a part of this work. It is shown through subjective testing that the proposed metric has a strong correlation with the Mean Opinion Score (MOS) for the perceived regularity of textures. The proposed method is also shown to be robust to geometric and photometric transformations and outperforms some of the popular texture regularity metrics in predicting the perceived regularity. The impact of the proposed metric to improve the performance of many image-processing applications is also presented. The influence of the perceived texture regularity on the perceptual quality of synthesized textures is demonstrated through building a synthesized textures database named SynTEX. It is shown through subjective testing that textures with different degrees of perceived regularities exhibit different degrees of vulnerability to artifacts resulting from different texture synthesis approaches. This work also proposes an algorithm for adaptively selecting the appropriate texture synthesis method based on the perceived regularity of the original texture. A reduced-reference texture quality metric for texture synthesis is also proposed as part of this work. The metric is based on the change in perceived regularity and the change in perceived granularity between the original and the synthesized textures. The perceived granularity is quantified through a new granularity metric that is proposed in this work. It is shown through subjective testing that the proposed quality metric, using just 2 parameters, has a strong correlation with the MOS for the fidelity of synthesized textures and outperforms the state-of-the-art full-reference quality metrics on 3 different texture databases. Finally, the ability of the proposed regularity metric in predicting the perceived degradation of textures due to compression and blur artifacts is also established.Dissertation/ThesisPh.D. Electrical Engineering 201

    Detecci贸n de situaciones de violencia f铆sica interpersonal en videos usando t茅cnicas de aprendizaje profundo

    Get PDF
    Dise帽a una arquitectura con el modelo de red neuronal convolucional Xception y LSTM para la detecci贸n de violencia f铆sica interpersonal en los videos de sistemas de vigilancia. Debido al aumento de inseguridad en el pa铆s y como medida preventiva, se busc贸 reforzar el sistema de videovigilancia, donde se enfoc贸 en la necesidad de integrar nuevas tecnolog铆as para supervisar la seguridad ciudadana como es el caso del uso de la visi贸n artificial. Para el entrenamiento, validaci贸n y prueba de la arquitectura del modelo propuesto, se utiliz贸 los conjuntos de datos Hockey Fight Dataset y Real Life Violence Situations Dataset. Los resultados obtenidos en la exactitud de nuestra propuesta en el conjunto de datos Hockey Fight Dataset supero a todos los dem谩s m茅todos. En el caso del conjunto de datos Real Life Violence Situations Dataset que cuenta 2000 videos en contraste de otros conjuntos de datos utilizados para la detecci贸n de violencia, se obtuvieron buenos resultados en la exactitud mayores al 90%.Per煤. Universidad Nacional Mayor de San Marcos. Vicerrectorado de Investigaci贸n y Posgrado. Proyectos de Investigaci贸n con Financiamiento para Grupos de Investigaci贸n. PCONFIGI. C贸digo: C21201361. Resoluci贸n: 005753-2021-R/UNMS

    Robust Mobile Visual Recognition System: From Bag of Visual Words to Deep Learning

    Get PDF
    With billions of images captured by mobile users everyday, automatically recognizing contents in such images has become a particularly important feature for various mobile apps, including augmented reality, product search, visual-based authentication etc. Traditionally, a client-server architecture is adopted such that the mobile client sends captured images/video frames to a cloud server, which runs a set of task-specific computer vision algorithms and sends back the recognition results. However, such scheme may cause problems related to user privacy, network stability/availability and device energy.In this dissertation, we investigate the problem of building a robust mobile visual recognition system that achieves high accuracy, low latency, low energy cost and privacy protection. Generally, we study two broad types of recognition methods: the bag of visual words (BOVW) based retrieval methods, which search the nearest neighbor image to a query image, and the state-of-the-art deep learning based methods, which recognize a given image using a trained deep neural network. The challenges of deploying BOVW based retrieval methods include: size of indexed image database, query latency, feature extraction efficiency and re-ranking performance. To address such challenges, we first proposed EMOD which enables efficient on-device image retrieval on a downloaded context-dependent partial image database. The efficiency is achieved by analyzing the BOVW processing pipeline and optimizing each module with algorithmic improvement.Recent deep learning based recognition approaches have been shown to greatly exceed the performance of traditional approaches. We identify several challenges of applying deep learning based recognition methods on mobile scenarios, namely energy efficiency and privacy protection for real-time visual processing, and mobile visual domain biases. Thus, we proposed two techniques to address them, (i) efficiently splitting the workload across heterogeneous computing resources, i.e., mobile devices and the cloud using our Moca framework, and (ii) using mobile visual domain adaptation as proposed in our collaborative edge-mediated platform DeepCham. Our extensive experiments on large-scale benchmark datasets and off-the-shelf mobile devices show our solutions provide better results than the state-of-the-art solutions
    corecore