184 research outputs found

    A performance analysis of dense stereo correspondence algorithms and error reduction techniques

    Get PDF
    Abstract: Dense stereo correspondence has been intensely studied and there exists a wide variety of proposed solutions in the literature. Different datasets have been constructed to test stereo algorithms, however, their ground truth formation and scene types vary. In this paper, state-of-the-art algorithms are compared using a number of datasets captured under varied conditions, with accuracy and density metrics forming the basis of a performance evaluation. Pre- and post-processing disparity map error reduction techniques are quantified

    Adaptive total variation deringing method for image interpolation

    Get PDF
    ABSTRACT This paper presents a new adaptive post-processing algorithm for ringing artifact reduction after image interpolation (upsampling). The algorithm is based on the concept of total variation (TV) for ringing control. It uses known TV of the blocks of the low-resolution image. Conditional gradient, subgradient and projection subgradient methods for this algorithm are considered and analyzed. A test set of 181300 overlapping 11x11 blocks of real images was used for local algorithm optimization and analysis. Local conditional gradient method shows the best objective and subjective results

    Image Quality Evaluation in Lossy Compressed Images

    Get PDF
    This research focuses on the quantification of image quality in lossy compressed images, exploring the impact of digital artefacts and scene characteristics upon image quality evaluation. A subjective paired comparison test was implemented to assess perceived quality of JPEG 2000 against baseline JPEG over a range of different scene types. Interval scales were generated for both algorithms, which indicated a subjective preference for JPEG 2000, particularly at low bit rates, and these were confirmed by an objective distortion measure. The subjective results did not follow this trend for some scenes however, and both algorithms were found to be scene dependent as a result of the artefacts produced at high compression rates. The scene dependencies were explored from the interval scale results, which allowed scenes to be grouped according to their susceptibilities to each of the algorithms. Groupings were correlated with scene measures applied in a linked study. A pilot study was undertaken to explore perceptibility thresholds of JPEG 2000 of the same set of images. This work was developed with a further experiment to investigate the thresholds of perceptibility and acceptability of higher resolution JPEG 2000 compressed images. A set of images was captured using a professional level full-frame Digital Single Lens Reflex camera, using a raw workflow and carefully controlled image-processing pipeline. The scenes were quantified using a set of simple scene metrics to classify them according to whether they were average, higher than, or lower than average, for a number of scene properties known to affect image compression and perceived image quality; these were used to make a final selection of test images. Image fidelity was investigated using the method of constant stimuli to quantify perceptibility thresholds and just noticeable differences (JNDs) of perceptibility. Thresholds and JNDs of acceptability were also quantified to explore suprathreshold quality evaluation. The relationships between the two thresholds were examined and correlated with the results from the scene measures, to identify more or less susceptible scenes. It was found that the level and differences between the two thresholds was an indicator of scene dependency and could be predicted by certain types of scene characteristics. A third study implemented the soft copy quality ruler as an alternative psychophysical method, by matching the quality of compressed images to a set of images varying in a single attribute, separated by known JND increments of quality. The imaging chain and image processing workflow were evaluated using objective measures of tone reproduction and spatial frequency response. An alternative approach to the creation of ruler images was implemented and tested, and the resulting quality rulers were used to evaluate a subset of the images from the previous study. The quality ruler was found to be successful in identifying scene susceptibilities and observer sensitivity. The fourth investigation explored the implementation of four different image quality metrics. These were the Modular Image Difference Metric, the Structural Similarity Metric, The Multi-scale Structural Similarity Metric and the Weighted Structural Similarity Metric. The metrics were tested against the subjective results and all were found to have linear correlation in terms of predictability of image quality

    Automatic lineament analysis techniques for remotely sensed imagery

    Get PDF
    Imperial Users onl

    Analysis of Hardware Accelerated Deep Learning and the Effects of Degradation on Performance

    Get PDF
    As convolutional neural networks become more prevalent in research and real world applications, the need for them to be faster and more robust will be a constant battle. This thesis investigates the effect of degradation being introduced to an image prior to object recognition with a convolutional neural network. As well as experimenting with methods to reduce the degradation and improve performance. Gaussian smoothing and additive Gaussian noise are both analyzed degradation models within this thesis and are reduced with Gaussian and Butterworth masks using unsharp masking and smoothing, respectively. The results show that each degradation is disruptive to the performance of YOLOv3, with Gaussian smoothing producing a mean average precision of less than 20% and Gaussian noise producing a mean average precision as low as 0%. Reduction methods applied to the data give results of 1%-21% mean average precision increase over the baseline, varying based on the degradation model. These methods are also applied to an 8-bit quantized implementation of YOLOv3, which is intended to run on a Xilinx ZCU104 FPGA, which showed to be as robust as the oating point network, with results within 2% mean average precision of the oating point network. With the ZCU104 being able to process images of 416x416 at 25 frames per second which is comparable to a NVIDIA 2080 RTX, FPGAs are a viable solution to computing object detection on the edge. In conclusion, this thesis shows that degradation causes performance of a convolutional neural network (quantized and oating point) to lose accuracy to a level that the network is unable to accurately predict objects. However, the degradation can be reduced, and in most cases can elevate the performance of the network by using computer vision techniques to reduce the noise within the image

    Techniques for enhancing digital images

    Get PDF
    The images obtain from either research studies or optical instruments are often corrupted with noise. Image denoising involves the manipulation of image data to produce a visually high quality image. This thesis reviews the existing denoising algorithms and the filtering approaches available for enhancing images and/or data transmission. Spatial-domain and Transform-domain digital image filtering algorithms have been used in the past to suppress different noise models. The different noise models can be either additive or multiplicative. Selection of the denoising algorithm is application dependent. It is necessary to have knowledge about the noise present in the image so as to select the appropriated denoising algorithm. Noise models may include Gaussian noise, Salt and Pepper noise, Speckle noise and Brownian noise. The Wavelet Transform is similar to the Fourier transform with a completely different merit function. The main difference between Wavelet transform and Fourier transform is that, in the Wavelet Transform, Wavelets are localized in both time and frequency. In the standard Fourier Transform, Wavelets are only localized in frequency. Wavelet analysis consists of breaking up the signal into shifted and scales versions of the original (or mother) Wavelet. The Wiener Filter (mean squared estimation error) finds implementations as a LMS filter (least mean squares), RLS filter (recursive least squares), or Kalman filter. Quantitative measure (metrics) of the comparison of the denoising algorithms is provided by calculating the Peak Signal to Noise Ratio (PSNR), the Mean Square Error (MSE) value and the Mean Absolute Error (MAE) evaluation factors. A combination of metrics including the PSNR, MSE, and MAE are often required to clearly assess the model performance
    • …
    corecore