24,045 research outputs found

    Self-assembly, modularity and physical complexity

    Get PDF
    We present a quantitative measure of physical complexity, based on the amount of information required to build a given physical structure through self-assembly. Our procedure can be adapted to any given geometry, and thus to any given type of physical system. We illustrate our approach using self-assembling polyominoes, and demonstrate the breadth of its potential applications by quantifying the physical complexity of molecules and protein complexes. This measure is particularly well suited for the detection of symmetry and modularity in the underlying structure, and allows for a quantitative definition of structural modularity. Furthermore we use our approach to show that symmetric and modular structures are favoured in biological self-assembly, for example of protein complexes. Lastly, we also introduce the notions of joint, mutual and conditional complexity, which provide a useful distance measure between physical structures.Comment: 9 pages, submitted for publicatio

    What is a quantum computer, and how do we build one?

    Full text link
    The DiVincenzo criteria for implementing a quantum computer have been seminal in focussing both experimental and theoretical research in quantum information processing. These criteria were formulated specifically for the circuit model of quantum computing. However, several new models for quantum computing (paradigms) have been proposed that do not seem to fit the criteria well. The question is therefore what are the general criteria for implementing quantum computers. To this end, a formal operational definition of a quantum computer is introduced. It is then shown that according to this definition a device is a quantum computer if it obeys the following four criteria: Any quantum computer must (1) have a quantum memory; (2) facilitate a controlled quantum evolution of the quantum memory; (3) include a method for cooling the quantum memory; and (4) provide a readout mechanism for subsets of the quantum memory. The criteria are met when the device is scalable and operates fault-tolerantly. We discuss various existing quantum computing paradigms, and how they fit within this framework. Finally, we lay out a roadmap for selecting an avenue towards building a quantum computer. This is summarized in a decision tree intended to help experimentalists determine the most natural paradigm given a particular physical implementation

    Minimally complex ion traps as modules for quantum communication and computing

    Full text link
    Optically linked ion traps are promising as components of network-based quantum technologies, including communication systems and modular computers. Experimental results achieved to date indicate that the fidelity of operations within each ion trap module will be far higher than the fidelity of operations involving the links; fortunately internal storage and processing can effectively upgrade the links through the process of purification. Here we perform the most detailed analysis to date on this purification task, using a protocol which is balanced to maximise fidelity while minimising the device complexity and the time cost of the process. Moreover we 'compile down' the quantum circuit to device-level operations including cooling and shutting events. We find that a linear trap with only five ions (two of one species, three of another) can support our protocol while incorporating desirable features such as 'global control', i.e. laser control pulses need only target an entire zone rather than differentiating one ion from its neighbour. To evaluate the capabilities of such a module we consider its use both as a universal communications node for quantum key distribution, and as the basic repeating unit of a quantum computer. For the latter case we evaluate the threshold for fault tolerant quantum computing using the surface code, finding acceptable fidelities for the 'raw' entangling link as low as 83% (or under 75% if an additional ion is available).Comment: 15 pages, 8 figure

    Network strategies for the new economy

    Get PDF
    In this paper we argue that the pace and scale of development in the information and communication technology industries (ICT) has had and continues to have major effects on the industry economics and competitive dynamics generally. We maintain that the size of changes in demand and supply conditions is forcing companies to make significant changes in the way they conceive and implement their strategies. We decompose the ICT industries into four levels, technology standards, supply chains, physical platforms, and consumer networks. The nature of these technologies and their cost characteristics coupled with higher degrees of knowledge specialisation is impelling companies to radical revisions of their attitudes towards cooperation and co-evolution with suppliers and customers. Where interdependencies between customers are particularly strong, we anticipate the possibility of winner-takes-all strategies. In these circumstances industry risks become very high and there will be significant consequences for competitive markets

    Towards Evaluating the Quality of a Spreadsheet: The Case of the Analytical Spreadsheet Model

    Get PDF
    We consider the challenge of creating guidelines to evaluate the quality of a spreadsheet model. We suggest four principles. First, state the domain-the spreadsheets to which the guidelines apply. Second, distinguish between the process by which a spreadsheet is constructed from the resulting spreadsheet artifact. Third, guidelines should be written in terms of the artifact, independent of the process. Fourth, the meaning of "quality" must be defined. We illustrate these principles with an example. We define the domain of "analytical spreadsheet models", which are used in business, finance, engineering, and science. We propose for discussion a framework and terminology for evaluating the quality of analytical spreadsheet models. This framework categorizes and generalizes the findings of previous work on the more narrow domain of financial spreadsheet models. We suggest that the ultimate goal is a set of guidelines for an evaluator, and a checklist for a developer.Comment: Proc. European Spreadsheet Risks Int. Grp. (EuSpRIG) 2011 ISBN 978-0-9566256-9-
    • …
    corecore