121,083 research outputs found

    A Multi-signal Variant for the GPU-based Parallelization of Growing Self-Organizing Networks

    Full text link
    Among the many possible approaches for the parallelization of self-organizing networks, and in particular of growing self-organizing networks, perhaps the most common one is producing an optimized, parallel implementation of the standard sequential algorithms reported in the literature. In this paper we explore an alternative approach, based on a new algorithm variant specifically designed to match the features of the large-scale, fine-grained parallelism of GPUs, in which multiple input signals are processed at once. Comparative tests have been performed, using both parallel and sequential implementations of the new algorithm variant, in particular for a growing self-organizing network that reconstructs surfaces from point clouds. The experimental results show that this approach allows harnessing in a more effective way the intrinsic parallelism that the self-organizing networks algorithms seem intuitively to suggest, obtaining better performances even with networks of smaller size.Comment: 17 page

    High-performance Kernel Machines with Implicit Distributed Optimization and Randomization

    Full text link
    In order to fully utilize "big data", it is often required to use "big models". Such models tend to grow with the complexity and size of the training data, and do not make strong parametric assumptions upfront on the nature of the underlying statistical dependencies. Kernel methods fit this need well, as they constitute a versatile and principled statistical methodology for solving a wide range of non-parametric modelling problems. However, their high computational costs (in storage and time) pose a significant barrier to their widespread adoption in big data applications. We propose an algorithmic framework and high-performance implementation for massive-scale training of kernel-based statistical models, based on combining two key technical ingredients: (i) distributed general purpose convex optimization, and (ii) the use of randomization to improve the scalability of kernel methods. Our approach is based on a block-splitting variant of the Alternating Directions Method of Multipliers, carefully reconfigured to handle very large random feature matrices, while exploiting hybrid parallelism typically found in modern clusters of multicore machines. Our implementation supports a variety of statistical learning tasks by enabling several loss functions, regularization schemes, kernels, and layers of randomized approximations for both dense and sparse datasets, in a highly extensible framework. We evaluate the ability of our framework to learn models on data from applications, and provide a comparison against existing sequential and parallel libraries.Comment: Work presented at MMDS 2014 (June 2014) and JSM 201

    An efficient genetic algorithm for large-scale planning of robust industrial wireless networks

    Get PDF
    An industrial indoor environment is harsh for wireless communications compared to an office environment, because the prevalent metal easily causes shadowing effects and affects the availability of an industrial wireless local area network (IWLAN). On the one hand, it is costly, time-consuming, and ineffective to perform trial-and-error manual deployment of wireless nodes. On the other hand, the existing wireless planning tools only focus on office environments such that it is hard to plan IWLANs due to the larger problem size and the deployed IWLANs are vulnerable to prevalent shadowing effects in harsh industrial indoor environments. To fill this gap, this paper proposes an overdimensioning model and a genetic algorithm based over-dimensioning (GAOD) algorithm for deploying large-scale robust IWLANs. As a progress beyond the state-of-the-art wireless planning, two full coverage layers are created. The second coverage layer serves as redundancy in case of shadowing. Meanwhile, the deployment cost is reduced by minimizing the number of access points (APs); the hard constraint of minimal inter-AP spatial paration avoids multiple APs covering the same area to be simultaneously shadowed by the same obstacle. The computation time and occupied memory are dedicatedly considered in the design of GAOD for large-scale optimization. A greedy heuristic based over-dimensioning (GHOD) algorithm and a random OD algorithm are taken as benchmarks. In two vehicle manufacturers with a small and large indoor environment, GAOD outperformed GHOD with up to 20% less APs, while GHOD outputted up to 25% less APs than a random OD algorithm. Furthermore, the effectiveness of this model and GAOD was experimentally validated with a real deployment system
    • …
    corecore