28 research outputs found

    A Comparative Study on Improvement of Image Compression Method using Hybrid DCT - DWT Techniques with Huffman Encoding for Wireless Sensor Network Application

    Get PDF
    Nowadays, the demands on the usage of wireless network are increasing rapidly from year to year. Wireless network is a large scale of area where many nodes are connecting to each other to communicate using a device. Primarily, wireless network also tend to be as a link to transmit and receive any multimedia such as image, sound, video, document and etc. In order to receive the transmitted media correctly, most type of media must be compressed before being transmitted and decompressed after being received by the device or else the device used must have the ability to read the media in a compressed way. In this paper, a hybrid compression of Discrete Cosine Transform (DCT) and Discrete Wavelet Transform (DWT) with Huffman encoding technique are proposed for Wireless Sensor Network (WSN) application. Data compression is very useful to remove the redundant data and reduce the size of image. After conducting a comprehensive observation, it is found that hybrid compression is suitable due to the process consist of the combination of multiple compression techniques which suits for Wireless Sensor Network’s application focusing on ZigBee platform

    Applications of Image-Based Computer Vision for Remote Surveillance of Slope Instability

    Get PDF
    Landslides and slope failures represent critical hazards for both the safety of local communities and the potential damage to economically relevant infrastructure such as roads, hydroelectric plants, pipelines, etc. Numerous surveillance methods, including ground-based radar, InSAR, Lidar, seismometers, and more recently computer vision, are available to monitor landslides and slope instability. However, the high cost, complexity, and intrinsic technical limitations of these methods frequently require the design of alternative and complementary techniques. Here, we provide an improved methodology for the application of image-based computer vision in landslide and rockfall monitoring. The newly developed open access Python-based software, Akh-Defo, uses optical flow velocity, image differencing and similarity index map techniques to calculate land deformation including landslides and rockfall. Akh-Defo is applied to two different datasets, notably ground- and satellite-based optical imagery for the Plinth Peak slope in British Columbia, Canada, and satellite optical imagery for the Mud Creek landslide in California, USA. Ground-based optical images were processed to evaluate the capability of Akh-Defo to identify rockfalls and measure land displacement in steep-slope terrains to complement LOS limitations of radar satellite images. Similarly, satellite optical images were processed to evaluate the capability of Akh-Defo to identify ground displacement in active landslide regions a few weeks to months prior to initiation of landslides. The Akh-Defo results were validated from two independent datasets including radar-imagery, processed using state of the art SqueeSAR algorithm for the Plinth Peak case study and very high-resolution temporal Lidar and photogrammetry digital surface elevation datasets for the Mud Creek case study. Our study shows that the Akh-Defo software complements InSAR by mitigating LOS limitations via processing ground-based optical imagery. Additionally, if applied to satellite optical imagery, it can be used as a first stage preliminary warning system (particularly when run on the cloud allowing near real-time processing) prior to processing more expensive but more accurate InSAR products such as SqueeSAR

    DERMATOLOGICAL IMAGE DENOISING USING ADAPTIVE HENLM METHOD

    Get PDF
    In this paper we propose automatic image denoising method based on Hermite functions (HeNLM). It is an extension of non-local means (NLM) algorithm. Differences between small image blocks (patches) are replaced by differences between feature vectors thus reducing computational complexity. The features are calculated in coordinate system connected with image gradient and are invariant to patch rotation. HeNLM method depends on the parameter that controls filtering strength. To chose automatically this parameter we use a no-reference denoising quality assessment method. It is based on Hessian matrix analysis. We compare the proposed method with full-reference methods using PSNR metrics, SSIM metrics, and its modifications MSSIM and CMSC. Image databases TID, DRIVE, BSD, and a set of dermatological immunofluorescence microscopy images were used for the tests. It was found that more perceptual CMSC and MSSIM metrics give worse correspondence than SSIM and PSNR to the results of information preservation by the non-reference image denoising

    Transpose Attack: Stealing Datasets with Bidirectional Training

    Full text link
    Deep neural networks are normally executed in the forward direction. However, in this work, we identify a vulnerability that enables models to be trained in both directions and on different tasks. Adversaries can exploit this capability to hide rogue models within seemingly legitimate models. In addition, in this work we show that neural networks can be taught to systematically memorize and retrieve specific samples from datasets. Together, these findings expose a novel method in which adversaries can exfiltrate datasets from protected learning environments under the guise of legitimate models. We focus on the data exfiltration attack and show that modern architectures can be used to secretly exfiltrate tens of thousands of samples with high fidelity, high enough to compromise data privacy and even train new models. Moreover, to mitigate this threat we propose a novel approach for detecting infected models.Comment: NDSS24 pape

    Perceptual Image Similarity Metrics and Applications.

    Full text link
    This dissertation presents research in perceptual image similarity metrics and applications, e.g., content-based image retrieval, perceptual image compression, image similarity assessment and texture analysis. The first part aims to design texture similarity metrics consistent with human perception. A new family of statistical texture similarity features, called Local Radius Index (LRI), and corresponding similarity metrics are proposed. Compared to state-of-the-art metrics in the STSIM family, LRI-based metrics achieve better texture retrieval performance with much less computation. When applied to the recently developed perceptual image coder, Matched Texture Coding (MTC), they enable similar performance while significantly accelerating encoding. Additionally, in photographic paper classification, LRI-based metrics also outperform pre-existing metrics. To fulfill the needs of texture classification and other applications, a rotation-invariant version of LRI, called Rotation-Invariant Local Radius Index (RI-LRI), is proposed. RI-LRI is also grayscale and illuminance insensitive. The corresponding similarity metric achieves texture classification accuracy comparable to state-of-the-art metrics. Moreover, its much lower dimensional feature vector requires substantially less computation and storage than other state-of-the-art texture features. The second part of the dissertation focuses on bilevel images, which are images whose pixels are either black or white. The contributions include new objective similarity metrics intended to quantify similarity consistent with human perception, and a subjective experiment to obtain ground truth for judging the performance of objective metrics. Several similarity metrics are proposed that outperform existing ones in the sense of attaining significantly higher Pearson and Spearman-rank correlations with the ground truth. The new metrics include Adjusted Percentage Error, Bilevel Gradient Histogram, Connected Components Comparison and combinations of such. Another portion of the dissertation focuses on the aforementioned MTC, which is a block-based image coder that uses texture similarity metrics to decide if blocks of the image can be encoded by pointing to perceptually similar ones in the already coded region. The key to its success is an effective texture similarity metric, such as an LRI-based metric, and an effective search strategy. Compared to traditional image compression algorithms, e.g., JPEG, MTC achieves similar coding rate with higher reconstruction quality. And the advantage of MTC becomes larger as coding rate decreases.PhDElectrical Engineering: SystemsUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/113586/1/yhzhai_1.pd

    Data Reduction Algorithms in Machine Learning and Data Science

    Get PDF
    Raw data are usually required to be pre-processed for better representation or discrimination of classes. This pre-processing can be done by data reduction, i.e., either reduction in dimensionality or numerosity (cardinality). Dimensionality reduction can be used for feature extraction or data visualization. Numerosity reduction is useful for ranking data points or finding the most and least important data points. This thesis proposes several algorithms for data reduction, known as dimensionality and numerosity reduction, in machine learning and data science. Dimensionality reduction tackles feature extraction and feature selection methods while numerosity reduction includes prototype selection and prototype generation approaches. This thesis focuses on feature extraction and prototype selection for data reduction. Dimensionality reduction methods can be divided into three categories, i.e., spectral, probabilistic, and neural network-based methods. The spectral methods have a geometrical point of view and are mostly reduced to the generalized eigenvalue problem. Probabilistic and network-based methods have stochastic and information theoretic foundations, respectively. Numerosity reduction methods can be divided into methods based on variance, geometry, and isolation. For dimensionality reduction, under the spectral category, I propose weighted Fisher discriminant analysis, Roweis discriminant analysis, and image quality aware embedding. I also propose quantile-quantile embedding as a probabilistic method where the distribution of embedding is chosen by the user. Backprojection, Fisher losses, and dynamic triplet sampling using Bayesian updating are other proposed methods in the neural network-based category. Backprojection is for training shallow networks with a projection-based perspective in manifold learning. Two Fisher losses are proposed for training Siamese triplet networks for increasing and decreasing the inter- and intra-class variances, respectively. Two dynamic triplet mining methods, which are based on Bayesian updating to draw triplet samples stochastically, are proposed. For numerosity reduction, principal sample analysis and instance ranking by matrix decomposition are the proposed variance-based methods; these methods rank instances using inter-/intra-class variances and matrix factorization, respectively. Curvature anomaly detection, in which the points are assumed to be the vertices of polyhedron, and isolation Mondrian forest are the proposed methods based on geometry and isolation, respectively. To assess the proposed tools developed for data reduction, I apply them to some applications in medical image analysis, image processing, and computer vision. Data reduction, used as a pre-processing tool, has different applications because it provides various ways of feature extraction and prototype selection for applying to different types of data. Dimensionality reduction extracts informative features and prototype selection selects the most informative data instances. For example, for medical image analysis, I use Fisher losses and dynamic triplet sampling for embedding histopathology image patches and demonstrating how different the tumorous cancer tissue types are from the normal ones. I also propose offline/online triplet mining using extreme distances for this embedding. In image processing and computer vision application, I propose Roweisfaces and Roweisposes for face recognition and 3D action recognition, respectively, using my proposed Roweis discriminant analysis method. I also introduce the concepts of anomaly landscape and anomaly path using the proposed curvature anomaly detection and use them to denoise images and video frames. I report extensive experiments, on different datasets, to show the effectiveness of the proposed algorithms. By experiments, I demonstrate that the proposed methods are useful for extracting informative features and instances for better accuracy, representation, prediction, class separation, data reduction, and embedding. I show that the proposed dimensionality reduction methods can extract informative features for better separation of classes. An example is obtaining an embedding space for separating cancer histopathology patches from the normal patches which helps hospitals diagnose cancers more easily in an automatic way. I also show that the proposed numerosity reduction methods are useful for ranking data instances based on their importance and reducing data volumes without a significant drop in performance of machine learning and data science algorithms

    Image Manipulation and Image Synthesis

    Get PDF
    Image manipulation is of historic importance. Ever since the advent of photography, pictures have been manipulated for various reasons. Historic rulers often used image manipulation techniques for the purpose of self-portrayal or propaganda. In many cases, the goal is to manipulate human behaviour by spreading credible misinformation. Photographs, by their nature, portray the real world and as such are more credible to humans. However, image manipulation may not only serve evil purposes. In this thesis, we propose and analyse methods for image manipulation that serve a positive purpose. Specifically, we treat image manipulation as a tool for solving other tasks. For this, we model image manipulation as an image-to-image translation (I2I) task, i.e., a system that receives an image as input and outputs a manipulated version of the input. We propose multiple I2I based methods. We demonstrate that I2I based image manipulation methods can be used to reduce motion blur in videos. Second, we show that I2I based image manipulation methods can be used for domain adaptation and domain extension. Specifically, we present a method that significantly improves the learning of semantic segmentation from synthetic source data. The same technique can be applied to learning nighttime semantic segmentation from daylight images. Next, we show that I2I can be used to enable weakly supervised object segmentation. We show that each individual task requires and allows for different levels of supervision during the training of deep models in order to achieve best performance. We discuss the importance of maintaining control over the output of such methods and show that, with reduced levels of supervision, methods for maintaining stability during training and for establishing control over the output of a system become increasingly important. We propose multiple methods that solve the issues that arise in such systems. Finally, we demonstrate that our proposed mechanisms for control can be adapted to synthesise images from scratch

    The Multiplicative Zak Transform, Dimension Reduction, and Wavelet Analysis of LIDAR Data

    Get PDF
    This thesis broadly introduces several techniques within the context of timescale analysis. The representation, compression and reconstruction of DEM and LIDAR data types is studied with directional wavelet methods and the wedgelet decomposition. The optimality of the contourlet transform, and then the wedgelet transform is evaluated with a valuable new structural similarity index. Dimension reduction for material classification is conducted with a frame-based kernel pipeline and a spectral-spatial method using wavelet packets. It is shown that these techniques can improve on baseline material classification methods while significantly reducing the amount of data. Finally, the multiplicative Zak transform is modified to allow the study and partial characterization of wavelet frames

    Compréhension de scènes urbaines par combinaison d'information 2D/3D

    Get PDF
    Cette thèse traite du problème de segmentation sémantique d'une séquence d'images calibrées acquises dans un environnement urbain. Ce problème consiste, plus précisément, à partitionner chaque image en régions représentant les objets de la scène (façades, routes, etc.). Ainsi, à chaque région est associée une étiquette sémantique. Dans notre approche, l'étiquetage s'opère via des primitives visuelles de niveau intermédiaire appelés super-pixels, lesquels regroupent des pixels similaires au sens de différents critères proposés dans la littérature, qu'ils soient photométriques (s'appuyant sur les couleurs) ou géométriques (limitant la taille des super-pixels formés). Contrairement à l'état de l'art, où les travaux récents traitant le même problème s'appuient en entrée sur une sur-segmentation initiale sans la remettre en cause, notre idée est de proposer, dans un contexte multi-vues, une nouvelle approche de constructeur de superpixels s'appuyant sur une analyse tridimensionnelle de la scène et, en particulier, de ses structures planes. Pour construire de «meilleurs» superpixels, une mesure de planéité locale, qui quantifie à quel point la zone traitée de l'image correspond à une surface plane de la scène, est introduite. Cette mesure est évaluée à partir d'une rectification homographique entre deux images proches, induites par un plan candidat au support des points 3D associés à la zone traitée. Nous analysons l'apport de la mesure UQI (Universal Quality Image) et montrons qu'elle se compare favorablement aux autres métriques qui ont le potentiel de détecter des structures planes. On introduit ensuite un nouvel algorithme de construction de super-pixels, fondé sur l'algorithme SLIC (Simple Linear Iterative Clustering) dont le principe est de regrouper les plus proches voisins au sens d'une distance fusionnant similarités en couleur et en distance, et qui intègre cette mesure de planéité. Ainsi la sur-segmentation obtenue, couplée à la cohérence interimages provenant de la validation de la contrainte de planéité locale de la scène, permet d'attribuer une étiquette à chaque entité et d'obtenir ainsi une segmentation sémantique qui partitionne l'image en objets plans
    corecore