6,146 research outputs found

    The Cost of Address Translation

    Full text link
    Modern computers are not random access machines (RAMs). They have a memory hierarchy, multiple cores, and virtual memory. In this paper, we address the computational cost of address translation in virtual memory. Starting point for our work is the observation that the analysis of some simple algorithms (random scan of an array, binary search, heapsort) in either the RAM model or the EM model (external memory model) does not correctly predict growth rates of actual running times. We propose the VAT model (virtual address translation) to account for the cost of address translations and analyze the algorithms mentioned above and others in the model. The predictions agree with the measurements. We also analyze the VAT-cost of cache-oblivious algorithms.Comment: A extended abstract of this paper was published in the proceedings of ALENEX13, New Orleans, US

    Automata with Nested Pebbles Capture First-Order Logic with Transitive Closure

    Get PDF
    String languages recognizable in (deterministic) log-space are characterized either by two-way (deterministic) multi-head automata, or following Immerman, by first-order logic with (deterministic) transitive closure. Here we elaborate this result, and match the number of heads to the arity of the transitive closure. More precisely, first-order logic with k-ary deterministic transitive closure has the same power as deterministic automata walking on their input with k heads, additionally using a finite set of nested pebbles. This result is valid for strings, ordered trees, and in general for families of graphs having a fixed automaton that can be used to traverse the nodes of each of the graphs in the family. Other examples of such families are grids, toruses, and rectangular mazes. For nondeterministic automata, the logic is restricted to positive occurrences of transitive closure. The special case of k=1 for trees, shows that single-head deterministic tree-walking automata with nested pebbles are characterized by first-order logic with unary deterministic transitive closure. This refines our earlier result that placed these automata between first-order and monadic second-order logic on trees.Comment: Paper for Logical Methods in Computer Science, 27 pages, 1 figur

    A statistical approach to the inverse problem in magnetoencephalography

    Full text link
    Magnetoencephalography (MEG) is an imaging technique used to measure the magnetic field outside the human head produced by the electrical activity inside the brain. The MEG inverse problem, identifying the location of the electrical sources from the magnetic signal measurements, is ill-posed, that is, there are an infinite number of mathematically correct solutions. Common source localization methods assume the source does not vary with time and do not provide estimates of the variability of the fitted model. Here, we reformulate the MEG inverse problem by considering time-varying locations for the sources and their electrical moments and we model their time evolution using a state space model. Based on our predictive model, we investigate the inverse problem by finding the posterior source distribution given the multiple channels of observations at each time rather than fitting fixed source parameters. Our new model is more realistic than common models and allows us to estimate the variation of the strength, orientation and position. We propose two new Monte Carlo methods based on sequential importance sampling. Unlike the usual MCMC sampling scheme, our new methods work in this situation without needing to tune a high-dimensional transition kernel which has a very high cost. The dimensionality of the unknown parameters is extremely large and the size of the data is even larger. We use Parallel Virtual Machine (PVM) to speed up the computation.Comment: Published in at http://dx.doi.org/10.1214/14-AOAS716 the Annals of Applied Statistics (http://www.imstat.org/aoas/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Hybrid Genetic Bees Algorithm applied to Single Machine Scheduling with Earliness and Tardiness Penalties

    Get PDF
    This paper presents a hybrid Genetic-Bees Algorithm based optimised solution for the single machine scheduling problem. The enhancement of the Bees Algorithm (BA) is conducted using the Genetic Algorithm's (GA's) operators during the global search stage. The proposed enhancement aims to increase the global search capability of the BA gradually with new additions. Although the BA has very successful implementations on various type of optimisation problems, it has found that the algorithm suffers from weak global search ability which increases the computational complexities on NP-hard type optimisation problems e.g. combinatorial/permutational type optimisation problems. This weakness occurs due to using a simple global random search operation during the search process. To reinforce the global search process in the BA, the proposed enhancement is utilised to increase exploration capability by expanding the number of fittest solutions through the genetical variations of promising solutions. The hybridisation process is realised by including two strategies into the basic BA, named as â\u80\u9creinforced global searchâ\u80\u9d and â\u80\u9cjumping functionâ\u80\u9d strategies. The reinforced global search strategy is the first stage of the hybridisation process and contains the mutation operator of the GA. The second strategy, jumping function strategy, consists of four GA operators as single point crossover, multipoint crossover, mutation and randomisation. To demonstrate the strength of the proposed solution, several experiments were carried out on 280 well-known single machine benchmark instances, and the results are presented by comparing to other well-known heuristic algorithms. According to the experiments, the proposed enhancements provides better capability to basic BA to jump from local minima, and GBA performed better compared to BA in terms of convergence and the quality of results. The convergence time reduced about 60% with about 30% better results for highly constrained jobs
    • …
    corecore