1,529 research outputs found

    An Extension of Slow Feature Analysis for Nonlinear Blind Source Separation

    Get PDF
    We present and test an extension of slow feature analysis as a novel approach to nonlinear blind source separation. The algorithm relies on temporal correlations and iteratively reconstructs a set of statistically independent sources from arbitrary nonlinear instantaneous mixtures. Simulations show that it is able to invert a complicated nonlinear mixture of two audio signals with a reliability of more than 9090\%. The algorithm is based on a mathematical analysis of slow feature analysis for the case of input data that are generated from statistically independent sources

    Toward single particle reconstruction without particle picking: Breaking the detection limit

    Full text link
    Single-particle cryo-electron microscopy (cryo-EM) has recently joined X-ray crystallography and NMR spectroscopy as a high-resolution structural method for biological macromolecules. In a cryo-EM experiment, the microscope produces images called micrographs. Projections of the molecule of interest are embedded in the micrographs at unknown locations, and under unknown viewing directions. Standard imaging techniques first locate these projections (detection) and then reconstruct the 3-D structure from them. Unfortunately, high noise levels hinder detection. When reliable detection is rendered impossible, the standard techniques fail. This is a problem especially for small molecules, which can be particularly hard to detect. In this paper, we propose a radically different approach: we contend that the structure could, in principle, be reconstructed directly from the micrographs, without intermediate detection. As a result, even small molecules should be within reach for cryo-EM. To support this claim, we setup a simplified mathematical model and demonstrate how our autocorrelation analysis technique allows to go directly from the micrographs to the sought signals. This involves only one pass over the micrographs, which is desirable for large experiments. We show numerical results and discuss challenges that lay ahead to turn this proof-of-concept into a competitive alternative to state-of-the-art algorithms

    Hybrid solutions to instantaneous MIMO blind separation and decoding: narrowband, QAM and square cases

    Get PDF
    Future wireless communication systems are desired to support high data rates and high quality transmission when considering the growing multimedia applications. Increasing the channel throughput leads to the multiple input and multiple output and blind equalization techniques in recent years. Thereby blind MIMO equalization has attracted a great interest.Both system performance and computational complexities play important roles in real time communications. Reducing the computational load and providing accurate performances are the main challenges in present systems. In this thesis, a hybrid method which can provide an affordable complexity with good performance for Blind Equalization in large constellation MIMO systems is proposed first. Saving computational cost happens both in the signal sep- aration part and in signal detection part. First, based on Quadrature amplitude modulation signal characteristics, an efficient and simple nonlinear function for the Independent Compo- nent Analysis is introduced. Second, using the idea of the sphere decoding, we choose the soft information of channels in a sphere, and overcome the so- called curse of dimensionality of the Expectation Maximization (EM) algorithm and enhance the final results simultaneously. Mathematically, we demonstrate in the digital communication cases, the EM algorithm shows Newton -like convergence.Despite the widespread use of forward -error coding (FEC), most multiple input multiple output (MIMO) blind channel estimation techniques ignore its presence, and instead make the sim- plifying assumption that the transmitted symbols are uncoded. However, FEC induces code structure in the transmitted sequence that can be exploited to improve blind MIMO channel estimates. In final part of this work, we exploit the iterative channel estimation and decoding performance for blind MIMO equalization. Experiments show the improvements achievable by exploiting the existence of coding structures and that it can access the performance of a BCJR equalizer with perfect channel information in a reasonable SNR range. All results are confirmed experimentally for the example of blind equalization in block fading MIMO systems

    Blind deconvolution of medical ultrasound images: parametric inverse filtering approach

    Get PDF
    ©2007 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or distribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE. This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.DOI: 10.1109/TIP.2007.910179The problem of reconstruction of ultrasound images by means of blind deconvolution has long been recognized as one of the central problems in medical ultrasound imaging. In this paper, this problem is addressed via proposing a blind deconvolution method which is innovative in several ways. In particular, the method is based on parametric inverse filtering, whose parameters are optimized using two-stage processing. At the first stage, some partial information on the point spread function is recovered. Subsequently, this information is used to explicitly constrain the spectral shape of the inverse filter. From this perspective, the proposed methodology can be viewed as a ldquohybridizationrdquo of two standard strategies in blind deconvolution, which are based on either concurrent or successive estimation of the point spread function and the image of interest. Moreover, evidence is provided that the ldquohybridrdquo approach can outperform the standard ones in a number of important practical cases. Additionally, the present study introduces a different approach to parameterizing the inverse filter. Specifically, we propose to model the inverse transfer function as a member of a principal shift-invariant subspace. It is shown that such a parameterization results in considerably more stable reconstructions as compared to standard parameterization methods. Finally, it is shown how the inverse filters designed in this way can be used to deconvolve the images in a nonblind manner so as to further improve their quality. The usefulness and practicability of all the introduced innovations are proven in a series of both in silico and in vivo experiments. Finally, it is shown that the proposed deconvolution algorithms are capable of improving the resolution of ultrasound images by factors of 2.24 or 6.52 (as judged by the autocorrelation criterion) depending on the type of regularization method used

    QRS Complex Separation from Convolutive Mixtures of Biolectrical Signals Acquired by Wearable Systems

    Get PDF
    Independent component analysis (ICA) has been widely used to remove artefacts from multichannel biomedical signal acquisitions under the hypothesis that there is statistical independence among the original sources. However, the basic ICA model does not take into account the influence on the mixing process of the different paths from the signal sources to the sensors In this study we propose a convolutive mixtures model in order to overcome the limitations of the basic ICA approach. The independent components are estimated in the frequency domain, where the convolutive model can be solved through an instantaneous mixing model. The signals are reconstructed back to the observation space resolving the ICA model ambiguities. Simulations are carried out to optimize of the proposed method for convolutive mixtures of electrocardiographic (ECG) and motion artefacts signals. The algorithm is tested on real ECG signals acquired by wearable systems in order to preserve the QRS complex when the signals are degraded by real life conditions of acquisition

    Improvement of speech recognition by nonlinear noise reduction

    Full text link
    The success of nonlinear noise reduction applied to a single channel recording of human voice is measured in terms of the recognition rate of a commercial speech recognition program in comparison to the optimal linear filter. The overall performance of the nonlinear method is shown to be superior. We hence demonstrate that an algorithm which has its roots in the theory of nonlinear deterministic dynamics possesses a large potential in a realistic application.Comment: see urbanowicz.org.p

    Complex-valued Adaptive Digital Signal Enhancement For Applications In Wireless Communication Systems

    Get PDF
    In recent decades, the wireless communication industry has attracted a great deal of research efforts to satisfy rigorous performance requirements and preserve high spectral efficiency. Along with this trend, I/Q modulation is frequently applied in modern wireless communications to develop high performance and high data rate systems. This has necessitated the need for applying efficient complex-valued signal processing techniques to highly-integrated, multi-standard receiver devices. In this dissertation, novel techniques for complex-valued digital signal enhancement are presented and analyzed for various applications in wireless communications. The first technique is a unified block processing approach to generate the complex-valued conjugate gradient Least Mean Square (LMS) techniques with optimal adaptations. The proposed algorithms exploit the concept of the complex conjugate gradients to find the orthogonal directions for updating the adaptive filter coefficients at each iteration. Along each orthogonal direction, the presented algorithms employ the complex Taylor series expansion to calculate time-varying convergence factors tailored for the adaptive filter coefficients. The performance of the developed technique is tested in the applications of channel estimation, channel equalization, and adaptive array beamforming. Comparing with the state of the art methods, the proposed techniques demonstrate improved performance and exhibit desirable characteristics for practical use. The second complex-valued signal processing technique is a novel Optimal Block Adaptive algorithm based on Circularity, OBA-C. The proposed OBA-C method compensates for a complex imbalanced signal by restoring its circularity. In addition, by utilizing the complex iv Taylor series expansion, the OBA-C method optimally updates the adaptive filter coefficients at each iteration. This algorithm can be applied to mitigate the frequency-dependent I/Q mismatch effects in analog front-end. Simulation results indicate that comparing with the existing methods, OBA-C exhibits superior convergence speed while maintaining excellent accuracy. The third technique is regarding interference rejection in communication systems. The research on both LMS and Independent Component Analysis (ICA) based techniques continues to receive significant attention in the area of interference cancellation. The performance of the LMS and ICA based approaches is studied for signals with different probabilistic distributions. Our research indicates that the ICA-based approach works better for super-Gaussian signals, while the LMS-based method is preferable for sub-Gaussian signals. Therefore, an appropriate choice of interference suppression algorithms can be made to satisfy the ever-increasing demand for better performance in modern receiver design

    Wireless Channel Equalization in Digital Communication Systems

    Get PDF
    Our modern society has transformed to an information-demanding system, seeking voice, video, and data in quantities that could not be imagined even a decade ago. The mobility of communicators has added more challenges. One of the new challenges is to conceive highly reliable and fast communication system unaffected by the problems caused in the multipath fading wireless channels. Our quest is to remove one of the obstacles in the way of achieving ultimately fast and reliable wireless digital communication, namely Inter-Symbol Interference (ISI), the intensity of which makes the channel noise inconsequential. The theoretical background for wireless channels modeling and adaptive signal processing are covered in first two chapters of dissertation. The approach of this thesis is not based on one methodology but several algorithms and configurations that are proposed and examined to fight the ISI problem. There are two main categories of channel equalization techniques, supervised (training) and blind unsupervised (blind) modes. We have studied the application of a new and specially modified neural network requiring very short training period for the proper channel equalization in supervised mode. The promising performance in the graphs for this network is presented in chapter 4. For blind modes two distinctive methodologies are presented and studied. Chapter 3 covers the concept of multiple cooperative algorithms for the cases of two and three cooperative algorithms. The select absolutely larger equalized signal and majority vote methods have been used in 2-and 3-algoirithm systems respectively. Many of the demonstrated results are encouraging for further research. Chapter 5 involves the application of general concept of simulated annealing in blind mode equalization. A limited strategy of constant annealing noise is experimented for testing the simple algorithms used in multiple systems. Convergence to local stationary points of the cost function in parameter space is clearly demonstrated and that justifies the use of additional noise. The capability of the adding the random noise to release the algorithm from the local traps is established in several cases
    • …
    corecore