151 research outputs found
Improving Unsupervised Defect Segmentation by Applying Structural Similarity to Autoencoders
Convolutional autoencoders have emerged as popular methods for unsupervised
defect segmentation on image data. Most commonly, this task is performed by
thresholding a pixel-wise reconstruction error based on an distance.
This procedure, however, leads to large residuals whenever the reconstruction
encompasses slight localization inaccuracies around edges. It also fails to
reveal defective regions that have been visually altered when intensity values
stay roughly consistent. We show that these problems prevent these approaches
from being applied to complex real-world scenarios and that it cannot be easily
avoided by employing more elaborate architectures such as variational or
feature matching autoencoders. We propose to use a perceptual loss function
based on structural similarity which examines inter-dependencies between local
image regions, taking into account luminance, contrast and structural
information, instead of simply comparing single pixel values. It achieves
significant performance gains on a challenging real-world dataset of
nanofibrous materials and a novel dataset of two woven fabrics over the state
of the art approaches for unsupervised defect segmentation that use pixel-wise
reconstruction error metrics
PET Synthesis via Self-supervised Adaptive Residual Estimation Generative Adversarial Network
Positron emission tomography (PET) is a widely used, highly sensitive
molecular imaging in clinical diagnosis. There is interest in reducing the
radiation exposure from PET but also maintaining adequate image quality. Recent
methods using convolutional neural networks (CNNs) to generate synthesized
high-quality PET images from low-dose counterparts have been reported to be
state-of-the-art for low-to-high image recovery methods. However, these methods
are prone to exhibiting discrepancies in texture and structure between
synthesized and real images. Furthermore, the distribution shift between
low-dose PET and standard PET has not been fully investigated. To address these
issues, we developed a self-supervised adaptive residual estimation generative
adversarial network (SS-AEGAN). We introduce (1) An adaptive residual
estimation mapping mechanism, AE-Net, designed to dynamically rectify the
preliminary synthesized PET images by taking the residual map between the
low-dose PET and synthesized output as the input, and (2) A self-supervised
pre-training strategy to enhance the feature representation of the coarse
generator. Our experiments with a public benchmark dataset of total-body PET
images show that SS-AEGAN consistently outperformed the state-of-the-art
synthesis methods with various dose reduction factors.Comment: This work has been submitted to the IEEE for possible publication.
Copyright may be transferred without notice, after which this version may no
longer be accessibl
Generative Adversarial Super-Resolution at the Edge with Knowledge Distillation
Single-Image Super-Resolution can support robotic tasks in environments where
a reliable visual stream is required to monitor the mission, handle
teleoperation or study relevant visual details. In this work, we propose an
efficient Generative Adversarial Network model for real-time Super-Resolution.
We adopt a tailored architecture of the original SRGAN and model quantization
to boost the execution on CPU and Edge TPU devices, achieving up to 200 fps
inference. We further optimize our model by distilling its knowledge to a
smaller version of the network and obtain remarkable improvements compared to
the standard training approach. Our experiments show that our fast and
lightweight model preserves considerably satisfying image quality compared to
heavier state-of-the-art models. Finally, we conduct experiments on image
transmission with bandwidth degradation to highlight the advantages of the
proposed system for mobile robotic applications
Interpolation of Low-Resolution Images for Improved Accuracy Using an ANN Quadratic Interpolator
The era of digital imaging has transitioned into a new one. Conversion to real-time, high-resolution images is considered vital. Interpolation is employed in order to increase the number of pixels per image, thereby enhancing spatial resolution. Interpolation's real advantage is that it can be deployed on user end devices. Despite raising the number of pixels per inch to enhances the spatial resolution, it may not improve the image's clarity, hence diminishing its quality. This strategy is designed to increase image quality by enhancing image sharpness and spatial resolution simultaneously. Proposed is an Artificial Neural Network (ANN) Quadratic Interpolator for interpolating 3-D images. This method applies Lagrange interpolating polynomial and Lagrange interpolating basis function to the parameter space using a deep neural network. The degree of the polynomial is determined by the frequency of gradient orientation events within the region of interest. By manipulating interpolation coefficients, images can be upscaled and enhanced. By mapping between low- and high-resolution images, the ANN quadratic interpolator optimizes the loss function. ANN Quadratic interpolator does a good work of reducing the amount of image artefacts that occur during the process of interpolation. The weights of the proposed ANN Quadratic interpolator are seeded by transfer learning, and the layers are trained, validated, and evaluated using a standard dataset. The proposed method outperforms a variety of cutting-edge picture interpolation algorithms.
- …