214,807 research outputs found

    Hydrographic Data Processing on a Robust, Network-coupled Parallel Cluster

    Get PDF
    There have been tremendous advances in acoustic sensor technologies and widespread adoption of multibeam echo-sounders in the recent past, which have enabled the efficient collection of large quantities of bathymetric data in every survey. However, timely dissemination of this data to the scientific community has been constrained by the relatively slow progress in the development of new data processing architectures. The current solutions for powerful, efficient and near-real time data processing systems entail high capital investments and technical complexities. Therefore, the installation base for these systems has been very small. The work presented here proposes a new architecture for bathymetric data processing based on parallel computing paradigms. The solution works by distributing the processing workload across a cluster of network-attached compute nodes. The success of using parallel processing for bathymetric data depends on the accurate measurement of the processing workload and its effective distribution across the compute nodes, thereby maximizing speedup and efficiency. These compute resources can be existing installations and other COTS components, such as blade servers, thereby reducing installation and maintenance expenditure. For workload determination, an estimation algorithm was developed that uses stochastic sampling of the raw bathymetric data file. This produces a low cost and high accuracy estimate of the processing requirements for each line to be processed. This workload information, coupled with file and system metadata, is used as input to different load balancing algorithms - First Come First Served (FCFS), Longest Job First (LJF) and Contention-Reduction (CR). The performance of FCFS and LJF algorithms is highly dependent on the characteristics of the input dataset while CR scheduling aims to characterize the input and adjust load distribution for the best combination of speedup and efficiency. The choice of these algorithms depends on the requirements of the installation, i.e. prioritization of speedup or efficiency. To ensure robustness, watchdog mechanisms monitor the state of all the components of the processing system and can react to system faults and failures, through a combination of automated and manual techniques. Although not part of the current implementation, there is potential for adding redundant critical components and to enable live-failover, thereby reducing or eliminating system downtime. The methods for workload estimation and distribution are templates for extending this framework to include additional types of bathymetric data and develop flexible, self-learning algorithms to deal with diverse datasets. This research lays the groundwork for the design of a ship-based system that would enable near-real time data processing and result in a faster ping-to-chart solution

    A Taxonomy of Workflow Management Systems for Grid Computing

    Full text link
    With the advent of Grid and application technologies, scientists and engineers are building more and more complex applications to manage and process large data sets, and execute scientific experiments on distributed resources. Such application scenarios require means for composing and executing complex workflows. Therefore, many efforts have been made towards the development of workflow management systems for Grid computing. In this paper, we propose a taxonomy that characterizes and classifies various approaches for building and executing workflows on Grids. We also survey several representative Grid workflow systems developed by various projects world-wide to demonstrate the comprehensiveness of the taxonomy. The taxonomy not only highlights the design and engineering similarities and differences of state-of-the-art in Grid workflow systems, but also identifies the areas that need further research.Comment: 29 pages, 15 figure

    Local Ensemble Transform Kalman Filter: a non-stationary control law for complex adaptive optics systems on ELTs

    Full text link
    We propose a new algorithm for an adaptive optics system control law which allows to reduce the computational burden in the case of an Extremely Large Telescope (ELT) and to deal with non-stationary behaviors of the turbulence. This approach, using Ensemble Transform Kalman Filter and localizations by domain decomposition is called the local ETKF: the pupil of the telescope is split up into various local domains and calculations for the update estimate of the turbulent phase on each domain are performed independently. This data assimilation scheme enables parallel computation of markedly less data during this update step. This adapts the Kalman Filter to large scale systems with a non-stationary turbulence model when the explicit storage and manipulation of extremely large covariance matrices are impossible. First simulation results are given in order to assess the theoretical analysis and to demonstrate the potentiality of this new control law for complex adaptive optics systems on ELTs.Comment: Proceedings of the AO4ELT3 conference; 8 pages, 3 figure

    An Asynchronous Parallel Approach to Sparse Recovery

    Full text link
    Asynchronous parallel computing and sparse recovery are two areas that have received recent interest. Asynchronous algorithms are often studied to solve optimization problems where the cost function takes the form ∑i=1Mfi(x)\sum_{i=1}^M f_i(x), with a common assumption that each fif_i is sparse; that is, each fif_i acts only on a small number of components of x∈Rnx\in\mathbb{R}^n. Sparse recovery problems, such as compressed sensing, can be formulated as optimization problems, however, the cost functions fif_i are dense with respect to the components of xx, and instead the signal xx is assumed to be sparse, meaning that it has only ss non-zeros where s≪ns\ll n. Here we address how one may use an asynchronous parallel architecture when the cost functions fif_i are not sparse in xx, but rather the signal xx is sparse. We propose an asynchronous parallel approach to sparse recovery via a stochastic greedy algorithm, where multiple processors asynchronously update a vector in shared memory containing information on the estimated signal support. We include numerical simulations that illustrate the potential benefits of our proposed asynchronous method.Comment: 5 pages, 2 figure
    • …
    corecore