151 research outputs found

    Improving Unsupervised Defect Segmentation by Applying Structural Similarity to Autoencoders

    Full text link
    Convolutional autoencoders have emerged as popular methods for unsupervised defect segmentation on image data. Most commonly, this task is performed by thresholding a pixel-wise reconstruction error based on an â„“p\ell^p distance. This procedure, however, leads to large residuals whenever the reconstruction encompasses slight localization inaccuracies around edges. It also fails to reveal defective regions that have been visually altered when intensity values stay roughly consistent. We show that these problems prevent these approaches from being applied to complex real-world scenarios and that it cannot be easily avoided by employing more elaborate architectures such as variational or feature matching autoencoders. We propose to use a perceptual loss function based on structural similarity which examines inter-dependencies between local image regions, taking into account luminance, contrast and structural information, instead of simply comparing single pixel values. It achieves significant performance gains on a challenging real-world dataset of nanofibrous materials and a novel dataset of two woven fabrics over the state of the art approaches for unsupervised defect segmentation that use pixel-wise reconstruction error metrics

    PET Synthesis via Self-supervised Adaptive Residual Estimation Generative Adversarial Network

    Full text link
    Positron emission tomography (PET) is a widely used, highly sensitive molecular imaging in clinical diagnosis. There is interest in reducing the radiation exposure from PET but also maintaining adequate image quality. Recent methods using convolutional neural networks (CNNs) to generate synthesized high-quality PET images from low-dose counterparts have been reported to be state-of-the-art for low-to-high image recovery methods. However, these methods are prone to exhibiting discrepancies in texture and structure between synthesized and real images. Furthermore, the distribution shift between low-dose PET and standard PET has not been fully investigated. To address these issues, we developed a self-supervised adaptive residual estimation generative adversarial network (SS-AEGAN). We introduce (1) An adaptive residual estimation mapping mechanism, AE-Net, designed to dynamically rectify the preliminary synthesized PET images by taking the residual map between the low-dose PET and synthesized output as the input, and (2) A self-supervised pre-training strategy to enhance the feature representation of the coarse generator. Our experiments with a public benchmark dataset of total-body PET images show that SS-AEGAN consistently outperformed the state-of-the-art synthesis methods with various dose reduction factors.Comment: This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessibl

    Generative Adversarial Super-Resolution at the Edge with Knowledge Distillation

    Get PDF
    Single-Image Super-Resolution can support robotic tasks in environments where a reliable visual stream is required to monitor the mission, handle teleoperation or study relevant visual details. In this work, we propose an efficient Generative Adversarial Network model for real-time Super-Resolution. We adopt a tailored architecture of the original SRGAN and model quantization to boost the execution on CPU and Edge TPU devices, achieving up to 200 fps inference. We further optimize our model by distilling its knowledge to a smaller version of the network and obtain remarkable improvements compared to the standard training approach. Our experiments show that our fast and lightweight model preserves considerably satisfying image quality compared to heavier state-of-the-art models. Finally, we conduct experiments on image transmission with bandwidth degradation to highlight the advantages of the proposed system for mobile robotic applications

    Interpolation of Low-Resolution Images for Improved Accuracy Using an ANN Quadratic Interpolator

    Get PDF
    The era of digital imaging has transitioned into a new one. Conversion to real-time, high-resolution images is considered vital. Interpolation is employed in order to increase the number of pixels per image, thereby enhancing spatial resolution. Interpolation's real advantage is that it can be deployed on user end devices. Despite raising the number of pixels per inch to enhances the spatial resolution, it may not improve the image's clarity, hence diminishing its quality. This strategy is designed to increase image quality by enhancing image sharpness and spatial resolution simultaneously. Proposed is an Artificial Neural Network (ANN) Quadratic Interpolator for interpolating 3-D images. This method applies Lagrange interpolating polynomial and Lagrange interpolating basis function to the parameter space using a deep neural network. The degree of the polynomial is determined by the frequency of gradient orientation events within the region of interest. By manipulating interpolation coefficients, images can be upscaled and enhanced. By mapping between low- and high-resolution images, the ANN quadratic interpolator optimizes the loss function. ANN Quadratic interpolator does a good work of reducing the amount of image artefacts that occur during the process of interpolation. The weights of the proposed ANN Quadratic interpolator are seeded by transfer learning, and the layers are trained, validated, and evaluated using a standard dataset. The proposed method outperforms a variety of cutting-edge picture interpolation algorithms.
    • …
    corecore