862 research outputs found

    Robust detail-preserving signal extraction

    Get PDF
    We discuss robust filtering procedures for signal extraction from noisy time series. Particular attention is paid to the preservation of relevant signal details like abrupt shifts. moving averages and running medians are widely used but have shortcomings when large spikes (outliers) or trends occur. Modifications like modified trimmed means and linear median hybrid filters combine advantages of both approaches, but they do not completely overcome the difficulties. Better solutions can be based on robust regression techniques, which even work in real time because of increased computational power and faster algorithms. Reviewing previous work we present filters for robust signal extraction and discuss their merits for preserving trends, abrupt shifts and local extremes as well as for the removal of outliers. --

    Mixed Noise Suppression in Color Images by Signal-Dependent LMS L-Filters

    Get PDF
    The paper is devoted to the signal-dependent (SD) design of adaptive LMS L-filters with marginal data ordering for color images. The same stem of SD processing of noised grayscale images was applied on noisy color images. Component-wise and multichannel modifications of SD LMS L-filter in R'G'B' (gamma corrected RGB signals) color space were developed. Both modifications for filtering two-dimensional static color images degraded by mixed noise consisting of additive Gaussian white noise and impulsive noise were used. Moreover, single-channel spatial impulse detectors as detectors of impulses and details were used, too. Considering experimental results, SD modifications of L-filters for noisy color images can be concluded to yield the best results

    On generalized adaptive neural filter

    Get PDF
    Linear filters have historically been used in the past as the most useful tools for suppressing noise in signal processing. It has been shown that the optimal filter which minimizes the mean square error (MSE) between the filter output and the desired output is a linear filter provided that the noise is additive white Gaussian noise (AWGN). However, in most signal processing applications, the noise in the channel through which a signal is transmitted is not AWGN; it is not stationary, and it may have unknown characteristics. To overcome the shortcomings of linear filters, nonlinear filters ranging from the median filters to stack filters have been developed. They have been successfully used in a number of applications, such as enhancing the signal-to-noise ratio of the telecommunication receivers, modeling the human vocal tract to synthesize speech in speech processing, and separating out the maternal and fetal electrocardiogram signals to diagnose prenatal ailments. In particular, stack filters have been shown to provide robust noise suppression, and are easily implementable in hardware, but configuring an optimal stack filter remains a challenge. This dissertation takes on this challenge by extending stack filters to a new class of nonlinear adaptive filters called generalized adaptive neural filters (GANFs). The objective of this work is to investigate their performance in terms of the mean absolute error criterion, to evaluate and predict the generalization of various discriminant functions employed for GANFs, and to address issues regarding their applications and implementation. It is shown that GANFs not only extend the class of stack filters, but also have better performance in terms of suppressing non-additive white Gaussian noise. Several results are drawn from the theoretical and experimental work: stack filters can be adaptively configured by neural networks; GANFs encompass a large class of nonlinear sliding-window filters which include stack filters; the mean absolute error (MAE) of the optimal GANF is upper-bounded by that of the optimal stack filter; a suitable class of discriminant functions can be determined before a training scheme is executed; VC dimension (VCdim) theory can be applied to determine the number of training samples; the algorithm presented in configuring GANFs is effective and robust

    Video enhancement : content classification and model selection

    Get PDF
    The purpose of video enhancement is to improve the subjective picture quality. The field of video enhancement includes a broad category of research topics, such as removing noise in the video, highlighting some specified features and improving the appearance or visibility of the video content. The common difficulty in this field is how to make images or videos more beautiful, or subjectively better. Traditional approaches involve lots of iterations between subjective assessment experiments and redesigns of algorithm improvements, which are very time consuming. Researchers have attempted to design a video quality metric to replace the subjective assessment, but so far it is not successful. As a way to avoid heuristics in the enhancement algorithm design, least mean square methods have received considerable attention. They can optimize filter coefficients automatically by minimizing the difference between processed videos and desired versions through a training. However, these methods are only optimal on average but not locally. To solve the problem, one can apply the least mean square optimization for individual categories that are classified by local image content. The most interesting example is Kondo’s concept of local content adaptivity for image interpolation, which we found could be generalized into an ideal framework for content adaptive video processing. We identify two parts in the concept, content classification and adaptive processing. By exploring new classifiers for the content classification and new models for the adaptive processing, we have generalized a framework for more enhancement applications. For the part of content classification, new classifiers have been proposed to classify different image degradations such as coding artifacts and focal blur. For the coding artifact, a novel classifier has been proposed based on the combination of local structure and contrast, which does not require coding block grid detection. For the focal blur, we have proposed a novel local blur estimation method based on edges, which does not require edge orientation detection and shows more robust blur estimation. With these classifiers, the proposed framework has been extended to coding artifact robust enhancement and blur dependant enhancement. With the content adaptivity to more image features, the number of content classes can increase significantly. We show that it is possible to reduce the number of classes without sacrificing much performance. For the part of model selection, we have introduced several nonlinear filters to the proposed framework. We have also proposed a new type of nonlinear filter, trained bilateral filter, which combines both advantages of the original bilateral filter and the least mean square optimization. With these nonlinear filters, the proposed framework show better performance than with linear filters. Furthermore, we have shown a proof-of-concept for a trained approach to obtain contrast enhancement by a supervised learning. The transfer curves are optimized based on the classification of global or local image content. It showed that it is possible to obtain the desired effect by learning from other computationally expensive enhancement algorithms or expert-tuned examples through the trained approach. Looking back, the thesis reveals a single versatile framework for video enhancement applications. It widens the application scope by including new content classifiers and new processing models and offers scalabilities with solutions to reduce the number of classes, which can greatly accelerate the algorithm design

    Adaptive weighted least squares algorithm for Volterra signal modeling

    No full text
    Published versio

    Iterative Multiscale Fusion and Night Vision Colorization of Multispectral Images

    Get PDF

    Development of Some Novel Nonlinear and Adaptive Digital Image Filters for Efficient Noise Suppression

    Get PDF
    Some nonlinear and adaptive digital image filtering algorithms have been developed in this thesis to suppress additive white Gaussian noise (AWGN), bipolar fixed-valued impulse, also called salt and pepper noise (SPN), random-valued impulse noise (RVIN) and their combinations quite effectively. The present state-of-art technology offers high quality sensors, cameras, electronic circuitry: application specific integrated circuits (ASIC), system on chip (SOC), etc., and high quality communication channels. Therefore, the noise level in images has been reduced drastically. In literature, many efficient nonlinear image filters are found that perform well under high noise conditions. But their performance is not so good under low noise conditions as compared to the extremely high computational complexity involved therein. Thus, it is felt that there is sufficient scope to investigate and develop quite efficient but simple algorithms to suppress low-power noise in an image. When..

    Novel proposal for prediction of CO2 course and occupancy recognition in Intelligent Buildings within IoT

    Get PDF
    Many direct and indirect methods, processes, and sensors available on the market today are used to monitor the occupancy of selected Intelligent Building (IB) premises and the living activities of IB residents. By recognizing the occupancy of individual spaces in IB, IB can be optimally automated in conjunction with energy savings. This article proposes a novel method of indirect occupancy monitoring using CO2, temperature, and relative humidity measured by means of standard operating measurements using the KNX (Konnex (standard EN 50090, ISO/IEC 14543)) technology to monitor laboratory room occupancy in an intelligent building within the Internet of Things (IoT). The article further describes the design and creation of a Software (SW) tool for ensuring connectivity of the KNX technology and the IoT IBM Watson platform in real-time for storing and visualization of the values measured using a Message Queuing Telemetry Transport (MQTT) protocol and data storage into a CouchDB type database. As part of the proposed occupancy determination method, the prediction of the course of CO2 concentration from the measured temperature and relative humidity values were performed using mathematical methods of Linear Regression, Neural Networks, and Random Tree (using IBM SPSS Modeler) with an accuracy higher than 90%. To increase the accuracy of the prediction, the application of suppression of additive noise from the CO2 signal predicted by CO2 using the Least mean squares (LMS) algorithm in adaptive filtering (AF) method was used within the newly designed method. In selected experiments, the prediction accuracy with LMS adaptive filtration was better than 95%.Web of Science1223art. no. 454

    Denoising techniques - a comparison

    Get PDF
    Visual information transmitted in the form of digital images is becoming a major method of communication in the modern age, but the image obtained after transmission is often corrupted with noise. The received image needs processing before it can be used in applications. Image denoising involves the manipulation of the image data to produce a visually high quality image. This thesis reviews the existing denoising algorithms, such as filtering approach, wavelet based approach, and multifractal approach, and performs their comparative study. Different noise models including additive and multiplicative types are used. They include Gaussian noise, salt and pepper noise, speckle noise and Brownian noise. Selection of the denoising algorithm is application dependent. Hence, it is necessary to have knowledge about the noise present in the image so as to select the appropriate denoising algorithm. The filtering approach has been proved to be the best when the image is corrupted with salt and pepper noise. The wavelet based approach finds applications in denoising images corrupted with Gaussian noise. In the case where the noise characteristics are complex, the multifractal approach can be used. A quantitative measure of comparison is provided by the signal to noise ratio of the image

    Fetal ECG Extraction from Maternal ECG using Attention-based CycleGAN

    Full text link
    Non-invasive fetal electrocardiogram (FECG) is used to monitor the electrical pulse of the fetal heart. Decomposing the FECG signal from maternal ECG (MECG) is a blind source separation problem, which is hard due to the low amplitude of FECG, the overlap of R waves, and the potential exposure to noise from different sources. Traditional decomposition techniques, such as adaptive filters, require tuning, alignment, or pre-configuration, such as modeling the noise or desired signal. to map MECG to FECG efficiently. The high correlation between maternal and fetal ECG parts decreases the performance of convolution layers. Therefore, the masking region of interest using the attention mechanism is performed for improving signal generators' precision. The sine activation function is also used since it could retain more details when converting two signal domains. Three available datasets from the Physionet, including A&D FECG, NI-FECG, and NI-FECG challenge, and one synthetic dataset using FECGSYN toolbox, are used to evaluate the performance. The proposed method could map abdominal MECG to scalp FECG with an average 98% R-Square [CI 95%: 97%, 99%] as the goodness of fit on A&D FECG dataset. Moreover, it achieved 99.7 % F1-score [CI 95%: 97.8-99.9], 99.6% F1-score [CI 95%: 98.2%, 99.9%] and 99.3% F1-score [CI 95%: 95.3%, 99.9%] for fetal QRS detection on, A&D FECG, NI-FECG and NI-FECG challenge datasets, respectively. These results are comparable to the state-of-the-art; thus, the proposed algorithm has the potential of being used for high-performance signal-to-signal conversion
    corecore