15 research outputs found

    An RGB-NIR Image Fusion Method for Improving Feature Matching

    Get PDF
    The quality of RGB images can be degraded by poor weather or lighting conditions. Thus, to make computer vision techniques work correctly, images need to be enhanced first. This paper proposes an RGB image enhancement method for improving feature matching which is a core step in most computer vision techniques. The proposed method decomposes near-infrared (NIR) image into fine detail, medium detail, and base images by using weighted least squares filters (WLSF) and boosts the medium detail image. Then, the fine and boosted medium detail images are combined, and the combined NIR detail image replaces the luminance detail image of an RGB image. Experiments demonstrates that the proposed method can effectively enhance RGB image; hence more stable image features are extracted. In addition, the method can minimize the loss of the useful visual (or optical) information of the original RGB image that can be used for other vision tasks

    Semantic Super-Resolution via Self-Distillation and Adversarial Learning

    No full text
    Semantic super-resolution (SR) is an approach that improves the SR performance by leveraging semantic information about the scene. This study develops a novel semantic SR method that is based on the generative adversarial network (GAN) framework and self-distillation. A discriminator is adversarially trained along with a generator to extract semantic features from images and distinguish semantic differences between images. To train the generator, an additional adversarial loss is computed from the discriminator’s outputs of SR images belonging to the same category and minimized via self-distillation. This guides the generator to learn implicit category-specific semantic priors. We conducted experiments for SR of text and face images using the Enhanced Deep Super-Resolution (EDSR) generator and the SRGAN discriminator. Experimental results showed that our method can contribute to improving both the quantitative and qualitative quality of SR images. Although the improvement varied depending on image category and dataset, the peak signal-to-noise ratio (PSNR) value increased by up to 0.87 dB and the perceptual index (PI) decreased by up to 0.17 by using our method. Our method outperformed existing semantic SR methods

    Semantic Super-Resolution of Text Images via Self-Distillation

    No full text
    This research develops an effective single-image super-resolution (SR) method that increases the resolution of scanned text or document images and improves their readability. To this end, we introduce a new semantic loss and propose a semantic SR method that guides an SR network to learn implicit text-specific semantic priors through self-distillation. Experiments on the enhanced deep SR (EDSR) model, one of the most popular SR networks, confirmed that semantic loss can contribute to further improving the quality of text SR images. Although the improvement varied depending on image resolution and dataset, the peak signal-to-noise ratio (PSNR) value was increased by up to 0.3 dB by introducing the semantic loss. The proposed method outperformed an existing semantic SR method

    No-Reference Blur Strength Estimation Based on Spectral Analysis of Blurred Images

    No full text

    High-Capacity Image Steganography Based on Side Match Block Matching

    No full text
    In this paper, we proposed an image steganography method which exploits a variant of block matching (BM) algorithm, called side match block matching (SMBM), and discrete wavelet transform (DWT) for embedding high-capacity images. The proposed method can greatly improve the embedding capacity by using SMBM. In addition, the proposed method resolves the problem with the loss of least significant bits in previous BM-based steganography methods and improves the confidentiality by embedding secret information in the DWT high frequency regions

    Rapid Generation of the State Codebook in Side Match Vector Quantization

    No full text

    Fast Feature Matching by Coarse-to-Fine Comparison of Rearranged SURF Descriptors

    No full text

    CNN-Based Ternary Classification for Image Steganalysis

    No full text
    This study proposes a convolutional neural network (CNN)-based steganalytic method that allows ternary classification to simultaneously identify WOW and UNIWARD, which are representative adaptive image steganographic algorithms. WOW and UNIWARD have very similar message embedding methods in terms of measuring and minimizing the degree of distortion of images caused by message embedding. This similarity between WOW and UNIWARD makes it difficult to distinguish between both algorithms even in a CNN-based classifier. Our experiments particularly show that WOW and UNIWARD cannot be distinguished by simply combining binary CNN-based classifiers learned to separately identify both algorithms. Therefore, to identify and classify WOW and UNIWARD, WOW and UNIWARD must be learned at the same time using a single CNN-based classifier designed for ternary classification. This study proposes a method for ternary classification that learns and classifies cover, WOW stego, and UNIWARD stego images using a single CNN-based classifier. A CNN structure and a preprocessing filter are also proposed to effectively classify/identify WOW and UNIWARD. Experiments using BOSSBase 1.01 database images confirmed that the proposed method could make a ternary classification with an accuracy of approximately 72%

    Improved Heart-Rate Measurement from Mobile Face Videos

    No full text
    Newtonian reaction to blood influx into the head at each heartbeat causes subtle head motion at the same frequency as the heartbeats. Thus, this head motion can be used to estimate the heart rate. Several studies have shown that heart rates can be measured accurately by tracking head motion using a desktop computer with a static camera. However, implementation of vision-based head motion tracking on smartphones demonstrated limited accuracy due to the hand-shaking problem caused by the non-static camera. The hand-shaking problem could not be handled effectively with only the frontal camera images. It also required a more accurate method to measure the periodicity of noisy signals. Therefore, this study proposes an improved head-motion-based heart-rate monitoring system using smartphones. To address the hand-shaking problem, the proposed system leverages the front and rear cameras available in most smartphones and dedicates each camera to tracking facial features that correspond to head motion and background features that correspond to hand-shaking. Then, the locations of facial features are adjusted using the average point of the background features. In addition, a correlation-based signal periodicity computation method is proposed to accurately separate the true heart-rate-related component from the head motion signal. The proposed system demonstrates improved accuracy (i.e., lower mean errors in heart-rate measurement) compared to conventional head-motion-based systems, and the accuracy is sufficient for daily heart-rate monitoring
    corecore