2,126 research outputs found

    Tradeoffs for language recognition on alternating machines

    Get PDF
    AbstractThe alternating machine having a separate input tape with k two-way, read-only heads, and a certain number of internal configurations, AM(k), is considered as a parallel computing model. For the complexity measure TIME · SPACE · PARALLELISM (TSP), the optimal lower bounds Ω(n2) and Ω(n3/2) respectively are proved for the recognition of specific languages on AM(1) and AM(k) respectively. For the complexity measure REVERSALS · SPACE · PARALLELISM (RSP), the lower bound Ω(n1/2) is established for the recognition of a specific language on AM(k). This result implies a polynomial lower bound on PARALLEL TIME · HARDWARE of parallel RAM's.Lower bounds on the complexity measures TIME · SPACE and REVERSALS · SPACE of nondeterministic machines are direct consequences of the result introduced above.All lower bounds obtained are substantially improved in the case that SPACE⩾ nɛ for 0<ɛ<1. Several strongest lower bounds for two-way and one-way alternating (deterministic, nondeterministic) multihead finite automata are obtained as direct consequences of these results. The hierarchies for the complexity measures TSP, RSP, TS and RS can be immediately achieved too

    High-performance Kernel Machines with Implicit Distributed Optimization and Randomization

    Full text link
    In order to fully utilize "big data", it is often required to use "big models". Such models tend to grow with the complexity and size of the training data, and do not make strong parametric assumptions upfront on the nature of the underlying statistical dependencies. Kernel methods fit this need well, as they constitute a versatile and principled statistical methodology for solving a wide range of non-parametric modelling problems. However, their high computational costs (in storage and time) pose a significant barrier to their widespread adoption in big data applications. We propose an algorithmic framework and high-performance implementation for massive-scale training of kernel-based statistical models, based on combining two key technical ingredients: (i) distributed general purpose convex optimization, and (ii) the use of randomization to improve the scalability of kernel methods. Our approach is based on a block-splitting variant of the Alternating Directions Method of Multipliers, carefully reconfigured to handle very large random feature matrices, while exploiting hybrid parallelism typically found in modern clusters of multicore machines. Our implementation supports a variety of statistical learning tasks by enabling several loss functions, regularization schemes, kernels, and layers of randomized approximations for both dense and sparse datasets, in a highly extensible framework. We evaluate the ability of our framework to learn models on data from applications, and provide a comparison against existing sequential and parallel libraries.Comment: Work presented at MMDS 2014 (June 2014) and JSM 201

    Reversals-space-parallelism tradeoffs for language recognition

    Get PDF

    Cumulative subject index volumes 52-55

    Get PDF

    Stochastic Optimization for Deep CCA via Nonlinear Orthogonal Iterations

    Full text link
    Deep CCA is a recently proposed deep neural network extension to the traditional canonical correlation analysis (CCA), and has been successful for multi-view representation learning in several domains. However, stochastic optimization of the deep CCA objective is not straightforward, because it does not decouple over training examples. Previous optimizers for deep CCA are either batch-based algorithms or stochastic optimization using large minibatches, which can have high memory consumption. In this paper, we tackle the problem of stochastic optimization for deep CCA with small minibatches, based on an iterative solution to the CCA objective, and show that we can achieve as good performance as previous optimizers and thus alleviate the memory requirement.Comment: in 2015 Annual Allerton Conference on Communication, Control and Computin
    corecore