139 research outputs found

    Methods for Signal Filtering and Modelling and Their Parallel Distributed Computing Implementation

    Get PDF
    In this thesis the problem of filtering and modelling one-dimensional discrete signals and implementation of corresponding parallel distributed algorithms will be addressed. In Chapter 2, the research areas of parallel distributed computing environments, rank-based nonlinear filter and fractal functions are reviewed. In Chapter 3, an Interactive Parallel Distributed Computing Environment (IPDCE) is implemented based on Parallel Virtual Machine (PVM) and an interactive application development tool, the Tc1 language. The approach we use is to provide a Tc1 version interface for all procedures of the PVM interface library so that users can utilize any PVM procedure to do their parallel computing interactively. In Chapter 4, an interactive parallel stack-filtering system is implemented, based on the IPDCE. The user can play with this filtering system in both traditional command mode and modern Graphics User Interface (GUI) mode. In order to reduce the time required to compute a standard stack filter, a new minimum threshold decomposition scheme is introduced and other techniques such as minimizing the number of logical operations and utilizing the CPU bit-fields parallel property are also suggested. In this filtering system the user can select sequential or parallel stack-filtering algorithms. The parallel distributed stack-filtering algorithm is implemented with equal task partitioning and PVM. Two numerical simulations show that the interactive parallel stack-filtering system is efficient for both the sequential and the parallel filtering algorithms. In Chapter 5, an extended Iterated Function System (IFS) interpolation method is introduced for modelling a given discrete signal. In order to get the solution of the inverse IFS problem in reasonable time, a suboptimal search algorithm, which estimates first the local self-affine region and then the map parameters is suggested, and the neighbourhood information of a self-affine region is used for enhancing the robustness of this suboptimal algorithm. The parallel distributed version of the in-verse IFS algorithm is implemented with equal task partitioning and using a Remote Procedure Call application programming interface library. The numerical simulation results show that the IFS approach achieves a higher signal to noise ratio than does an existing approach based on autoregressive modelling for self-affine and approximately self-affine one-dimensional signals and, when the number of computers is small, the speed-up ratio is almost linear. In Chapter 6, inverse IFS interpolation is introduced to model self-affine and approximately self-affine one-dimensional signals corrupted by Gaussian noise. Local cross-validation is applied for compromising between the degree of smoothness and fidelity to the data. The parallel distributed version of the inverse algorithm is implemented in Parallel Virtual Machine (PVM) with static optimal task partitioning. A simple computing model is applied which partitions tasks based on only each computer's capability. Several numerical simulation results show that the new IFS inverse algorithm achieves a higher signal to noise ratio than does existing autoregressive modelling for noisy self-affine or approximately self-affine signals.- There is little machine idle time relative to computing time in the optimal task partition mode. In Chapter 7, local IFS interpolation, which realises the IFS limit for self-affine data, is applied to model non self-affi.ne signals. It is difficult, however, to explore the whole parameter space to achieve globally optimal parameter estimation. A two-stage search scheme is suggested to estimate the self-affine region and the associated region parameters so that a suboptimal solution can be obtained in reasonable time. In the first stage, we calculate the self-affine region under the condition that the associated region length is twice that of the self-affine region. Then the second stage calculates the associated region for each self-affine region using a full search space. In order to combat the performance degradation caused by the the difference of machines capabilities and unpredictable external loads, a dynamic load-balance technique based on a data parallelism scheme is applied in the parallel distributed version of the inverse local IFS algorithm. Some numerical simulations show that our inverse local IFS algorithm works efficiently for several types of one-dimensional signal, and the parallel version with dynamic load balance can automatically ensure that each machine is busy with computing and with low idle time

    Efficient Multiband Algorithms for Blind Source Separation

    Get PDF
    The problem of blind separation refers to recovering original signals, called source signals, from the mixed signals, called observation signals, in a reverberant environment. The mixture is a function of a sequence of original speech signals mixed in a reverberant room. The objective is to separate mixed signals to obtain the original signals without degradation and without prior information of the features of the sources. The strategy used to achieve this objective is to use multiple bands that work at a lower rate, have less computational cost and a quicker convergence than the conventional scheme. Our motivation is the competitive results of unequal-passbands scheme applications, in terms of the convergence speed. The objective of this research is to improve unequal-passbands schemes by improving the speed of convergence and reducing the computational cost. The first proposed work is a novel maximally decimated unequal-passbands scheme.This scheme uses multiple bands that make it work at a reduced sampling rate, and low computational cost. An adaptation approach is derived with an adaptation step that improved the convergence speed. The performance of the proposed scheme was measured in different ways. First, the mean square errors of various bands are measured and the results are compared to a maximally decimated equal-passbands scheme, which is currently the best performing method. The results show that the proposed scheme has a faster convergence rate than the maximally decimated equal-passbands scheme. Second, when the scheme is tested for white and coloured inputs using a low number of bands, it does not yield good results; but when the number of bands is increased, the speed of convergence is enhanced. Third, the scheme is tested for quick changes. It is shown that the performance of the proposed scheme is similar to that of the equal-passbands scheme. Fourth, the scheme is also tested in a stationary state. The experimental results confirm the theoretical work. For more challenging scenarios, an unequal-passbands scheme with over-sampled decimation is proposed; the greater number of bands, the more efficient the separation. The results are compared to the currently best performing method. Second, an experimental comparison is made between the proposed multiband scheme and the conventional scheme. The results show that the convergence speed and the signal-to-interference ratio of the proposed scheme are higher than that of the conventional scheme, and the computation cost is lower than that of the conventional scheme

    High-performance hardware accelerators for image processing in space applications

    Get PDF
    Mars is a hard place to reach. While there have been many notable success stories in getting probes to the Red Planet, the historical record is full of bad news. The success rate for actually landing on the Martian surface is even worse, roughly 30%. This low success rate must be mainly credited to the Mars environment characteristics. In the Mars atmosphere strong winds frequently breath. This phenomena usually modifies the lander descending trajectory diverging it from the target one. Moreover, the Mars surface is not the best place where performing a safe land. It is pitched by many and close craters and huge stones, and characterized by huge mountains and hills (e.g., Olympus Mons is 648 km in diameter and 27 km tall). For these reasons a mission failure due to a landing in huge craters, on big stones or on part of the surface characterized by a high slope is highly probable. In the last years, all space agencies have increased their research efforts in order to enhance the success rate of Mars missions. In particular, the two hottest research topics are: the active debris removal and the guided landing on Mars. The former aims at finding new methods to remove space debris exploiting unmanned spacecrafts. These must be able to autonomously: detect a debris, analyses it, in order to extract its characteristics in terms of weight, speed and dimension, and, eventually, rendezvous with it. In order to perform these tasks, the spacecraft must have high vision capabilities. In other words, it must be able to take pictures and process them with very complex image processing algorithms in order to detect, track and analyse the debris. The latter aims at increasing the landing point precision (i.e., landing ellipse) on Mars. Future space-missions will increasingly adopt Video Based Navigation systems to assist the entry, descent and landing (EDL) phase of space modules (e.g., spacecrafts), enhancing the precision of automatic EDL navigation systems. For instance, recent space exploration missions, e.g., Spirity, Oppurtunity, and Curiosity, made use of an EDL procedure aiming at following a fixed and precomputed descending trajectory to reach a precise landing point. This approach guarantees a maximum landing point precision of 20 km. By comparing this data with the Mars environment characteristics, it is possible to understand how the mission failure probability still remains really high. A very challenging problem is to design an autonomous-guided EDL system able to even more reduce the landing ellipse, guaranteeing to avoid the landing in dangerous area of Mars surface (e.g., huge craters or big stones) that could lead to the mission failure. The autonomous behaviour of the system is mandatory since a manual driven approach is not feasible due to the distance between Earth and Mars. Since this distance varies from 56 to 100 million of km approximately due to the orbit eccentricity, even if a signal transmission at the light speed could be possible, in the best case the transmission time would be around 31 minutes, exceeding so the overall duration of the EDL phase. In both applications, algorithms must guarantee self-adaptability to the environmental conditions. Since the Mars (and in general the space) harsh conditions are difficult to be predicted at design time, these algorithms must be able to automatically tune the internal parameters depending on the current conditions. Moreover, real-time performances are another key factor. Since a software implementation of these computational intensive tasks cannot reach the required performances, these algorithms must be accelerated via hardware. For this reasons, this thesis presents my research work done on advanced image processing algorithms for space applications and the associated hardware accelerators. My research activity has been focused on both the algorithm and their hardware implementations. Concerning the first aspect, I mainly focused my research effort to integrate self-adaptability features in the existing algorithms. While concerning the second, I studied and validated a methodology to efficiently develop, verify and validate hardware components aimed at accelerating video-based applications. This approach allowed me to develop and test high performance hardware accelerators that strongly overcome the performances of the actual state-of-the-art implementations. The thesis is organized in four main chapters. Chapter 2 starts with a brief introduction about the story of digital image processing. The main content of this chapter is the description of space missions in which digital image processing has a key role. A major effort has been spent on the missions in which my research activity has a substantial impact. In particular, for these missions, this chapter deeply analizes and evaluates the state-of-the-art approaches and algorithms. Chapter 3 analyzes and compares the two technologies used to implement high performances hardware accelerators, i.e., Application Specific Integrated Circuits (ASICs) and Field Programmable Gate Arrays (FPGAs). Thanks to this information the reader may understand the main reasons behind the decision of space agencies to exploit FPGAs instead of ASICs for high-performance hardware accelerators in space missions, even if FPGAs are more sensible to Single Event Upsets (i.e., transient error induced on hardware component by alpha particles and solar radiation in space). Moreover, this chapter deeply describes the three available space-grade FPGA technologies (i.e., One-time Programmable, Flash-based, and SRAM-based), and the main fault-mitigation techniques against SEUs that are mandatory for employing space-grade FPGAs in actual missions. Chapter 4 describes one of the main contribution of my research work: a library of high-performance hardware accelerators for image processing in space applications. The basic idea behind this library is to offer to designers a set of validated hardware components able to strongly speed up the basic image processing operations commonly used in an image processing chain. In other words, these components can be directly used as elementary building blocks to easily create a complex image processing system, without wasting time in the debug and validation phase. This library groups the proposed hardware accelerators in IP-core families. The components contained in a same family share the same provided functionality and input/output interface. This harmonization in the I/O interface enables to substitute, inside a complex image processing system, components of the same family without requiring modifications to the system communication infrastructure. In addition to the analysis of the internal architecture of the proposed components, another important aspect of this chapter is the methodology used to develop, verify and validate the proposed high performance image processing hardware accelerators. This methodology involves the usage of different programming and hardware description languages in order to support the designer from the algorithm modelling up to the hardware implementation and validation. Chapter 5 presents the proposed complex image processing systems. In particular, it exploits a set of actual case studies, associated with the most recent space agency needs, to show how the hardware accelerator components can be assembled to build a complex image processing system. In addition to the hardware accelerators contained in the library, the described complex system embeds innovative ad-hoc hardware components and software routines able to provide high performance and self-adaptable image processing functionalities. To prove the benefits of the proposed methodology, each case study is concluded with a comparison with the current state-of-the-art implementations, highlighting the benefits in terms of performances and self-adaptability to the environmental conditions

    Blind source separation via independent and sparse component analysis with application to temporomandibular disorder

    Get PDF
    Blind source separation (BSS) addresses the problem of separating multi channel signals observed by generally spatially separated sensors into their constituent underlying sources. The passage of these sources through an unknown mixing medium results in these observed multichannel signals. This study focuses on BSS, with special emphasis on its application to the temporomandibular joint disorder (TMD). TMD refers to all medical problems related to the temporomandibular joint (TMJ), which holds the lower jaw (mandible) and the temporal bone (skull). The overall objective of the work is to extract the two TMJ sound sources generated by the two TMJs, from the bilateral recordings obtained from the auditory canals, so as to aid the clinician in diagnosis and planning treatment policies. Firstly, the concept of 'variable tap length' is adopted in convolutive blind source separation. This relatively new concept has attracted attention in the field of adaptive signal processing, notably the least mean square (LMS) algorithm, but has not yet been introduced in the context of blind signal separation. The flexibility of the tap length of the proposed approach allows for the optimum tap length to be found, thereby mitigating computational complexity or catering for fractional delays arising in source separation. Secondly, a novel fixed point BSS algorithm based on Ferrante's affine transformation is proposed. Ferrante's affine transformation provides the freedom to select the eigenvalues of the Jacobian matrix of the fixed point function and thereby improves the convergence properties of the fixed point iteration. Simulation studies demonstrate the improved convergence of the proposed approach compared to the well-known fixed point FastICA algorithm. Thirdly, the underdetermined blind source separation problem using a filtering approach is addressed. An extension of the FastICA algorithm is devised which exploits the disparity in the kurtoses of the underlying sources to estimate the mixing matrix and thereafter achieves source recovery by employing the i-norm algorithm. Additionally, it will be shown that FastICA can also be utilised to extract the sources. Furthermore, it is illustrated how this scenario is particularly suitable for the separation of TMJ sounds. Finally, estimation of fractional delays between the mixtures of the TMJ sources is proposed as a means for TMJ separation. The estimation of fractional delays is shown to simplify the source separation to a case of in stantaneous BSS. Then, the estimated delay allows for an alignment of the TMJ mixtures, thereby overcoming a spacing constraint imposed by a well- known BSS technique, notably the DUET algorithm. The delay found from the TMJ bilateral recordings corroborates with the range reported in the literature. Furthermore, TMJ source localisation is also addressed as an aid to the dental specialist.EThOS - Electronic Theses Online ServiceGBUnited Kingdo
    • …
    corecore