47,890 research outputs found

    On space-bounded synchronized alternating Turing machines

    Get PDF
    AbstractWe continue the study of the computational power of synchronized alternating Turing machines (SATM) introduced in (Hromkovič 1986, Slobodová 1987, 1988a, b) to allow communication via synchronization among processes of alternating Turing machines. We are interested in comparing the four main classes of space-bounded synchronized alternating Turing machines obtained by adding or removing off-line capability and nondeterminism (1SUTM(S(n)), SUTM(S(n)), 1SATM(S(n)), and SATM(S(n)) against one another and against other variants of alternating Turing machines. Denoting the class of languages accepted by machines in C by L(C), we show as our main results that L(1SUTM(S(n))) ⊂ L(SUTM(S(n))) ⊂ L(1SATM(S(n)))= L(SATM(S(n))) for all space-bounded functions S(n)ϵo(n), and L(1SUTM(S(n)))= L(SUTM(S(n))) ⊂ L(1SATM(S(n)))=L(SATM(S(n))) for S(n)) ⩾ n. Furthermore, we show that for log log(n) ⩽ S(n)ϵo(log(n)), L(1SUTM(S(n))) is incomparable to L[1] ATM(S(n))). L(UTM(S(n))), L(1MUTM(S(n))), and L(MUTM(S(n))), where MATMs are alternating Turing machines with modified acceptance proposed in (Inoue 1989); in contrast, we show that these relationships become proper inclusions when log(n) ⩽ S(n)ϵo(n).For deterministic synchronized alternating finite automata with at most k processes (1DSA(k)FA and DSA(k)FA) we establish a tight hierarchy on the number of processes for the one-way case, namely, L(1DSA(n)FA) ⊂ L(1DSA(n+1)FA) for all n > 0, and show that L(1DFA(2)) − ∪k=1∞L(DSA(k)FA) ≠ ∅, where DFA(k) denotes deterministic k-head finite automata. Finally we investigate closure properties under Boolean operations for some of these classes of languages

    D-ADMM: A Communication-Efficient Distributed Algorithm For Separable Optimization

    Full text link
    We propose a distributed algorithm, named Distributed Alternating Direction Method of Multipliers (D-ADMM), for solving separable optimization problems in networks of interconnected nodes or agents. In a separable optimization problem there is a private cost function and a private constraint set at each node. The goal is to minimize the sum of all the cost functions, constraining the solution to be in the intersection of all the constraint sets. D-ADMM is proven to converge when the network is bipartite or when all the functions are strongly convex, although in practice, convergence is observed even when these conditions are not met. We use D-ADMM to solve the following problems from signal processing and control: average consensus, compressed sensing, and support vector machines. Our simulations show that D-ADMM requires less communications than state-of-the-art algorithms to achieve a given accuracy level. Algorithms with low communication requirements are important, for example, in sensor networks, where sensors are typically battery-operated and communicating is the most energy consuming operation.Comment: To appear in IEEE Transactions on Signal Processin

    Structural Drift: The Population Dynamics of Sequential Learning

    Get PDF
    We introduce a theory of sequential causal inference in which learners in a chain estimate a structural model from their upstream teacher and then pass samples from the model to their downstream student. It extends the population dynamics of genetic drift, recasting Kimura's selectively neutral theory as a special case of a generalized drift process using structured populations with memory. We examine the diffusion and fixation properties of several drift processes and propose applications to learning, inference, and evolution. We also demonstrate how the organization of drift process space controls fidelity, facilitates innovations, and leads to information loss in sequential learning with and without memory.Comment: 15 pages, 9 figures; http://csc.ucdavis.edu/~cmg/compmech/pubs/sdrift.ht
    corecore