135 research outputs found

    26. Theorietag Automaten und Formale Sprachen 23. Jahrestagung Logik in der Informatik: Tagungsband

    Get PDF
    Der Theorietag ist die Jahrestagung der Fachgruppe Automaten und Formale Sprachen der Gesellschaft für Informatik und fand erstmals 1991 in Magdeburg statt. Seit dem Jahr 1996 wird der Theorietag von einem eintägigen Workshop mit eingeladenen Vorträgen begleitet. Die Jahrestagung der Fachgruppe Logik in der Informatik der Gesellschaft für Informatik fand erstmals 1993 in Leipzig statt. Im Laufe beider Jahrestagungen finden auch die jährliche Fachgruppensitzungen statt. In diesem Jahr wird der Theorietag der Fachgruppe Automaten und Formale Sprachen erstmalig zusammen mit der Jahrestagung der Fachgruppe Logik in der Informatik abgehalten. Organisiert wurde die gemeinsame Veranstaltung von der Arbeitsgruppe Zuverlässige Systeme des Instituts für Informatik an der Christian-Albrechts-Universität Kiel vom 4. bis 7. Oktober im Tagungshotel Tannenfelde bei Neumünster. Während des Tre↵ens wird ein Workshop für alle Interessierten statt finden. In Tannenfelde werden • Christoph Löding (Aachen) • Tomás Masopust (Dresden) • Henning Schnoor (Kiel) • Nicole Schweikardt (Berlin) • Georg Zetzsche (Paris) eingeladene Vorträge zu ihrer aktuellen Arbeit halten. Darüber hinaus werden 26 Vorträge von Teilnehmern und Teilnehmerinnen gehalten, 17 auf dem Theorietag Automaten und formale Sprachen und neun auf der Jahrestagung Logik in der Informatik. Der vorliegende Band enthält Kurzfassungen aller Beiträge. Wir danken der Gesellschaft für Informatik, der Christian-Albrechts-Universität zu Kiel und dem Tagungshotel Tannenfelde für die Unterstützung dieses Theorietags. Ein besonderer Dank geht an das Organisationsteam: Maike Bradler, Philipp Sieweck, Joel Day. Kiel, Oktober 2016 Florin Manea, Dirk Nowotka und Thomas Wilk

    Evolving Interesting Initial Conditions for Cellular Automata of the Game of Life Type

    Full text link
    The use of a genetic algorithm to obtain "interesting" initial conditions for cellular automata of the family of Conway's Game of Life is described in this paper. The conditions have been selected so as to maximize the number of gliders, R-pentominoes, and similar structures generated during the execution of the automata. Besides the original Game of Life rules, we have tested automata with similar rules, such as HighLife and B38S23, as well as mixed and time-dependent rules. We have concluded that the temporal invariance of the rules of these automata does not seem to be a requirement for the existence of the selected structures

    Cellular Automata Models of Road Traffic

    Full text link
    In this paper, we give an elaborate and understandable review of traffic cellular automata (TCA) models, which are a class of computationally efficient microscopic traffic flow models. TCA models arise from the physics discipline of statistical mechanics, having the goal of reproducing the correct macroscopic behaviour based on a minimal description of microscopic interactions. After giving an overview of cellular automata (CA) models, their background and physical setup, we introduce the mathematical notations, show how to perform measurements on a TCA model's lattice of cells, as well as how to convert these quantities into real-world units and vice versa. The majority of this paper then relays an extensive account of the behavioural aspects of several TCA models encountered in literature. Already, several reviews of TCA models exist, but none of them consider all the models exclusively from the behavioural point of view. In this respect, our overview fills this void, as it focusses on the behaviour of the TCA models, by means of time-space and phase-space diagrams, and histograms showing the distributions of vehicles' speeds, space, and time gaps. In the report, we subsequently give a concise overview of TCA models that are employed in a multi-lane setting, and some of the TCA models used to describe city traffic as a two-dimensional grid of cells, or as a road network with explicitly modelled intersections. The final part of the paper illustrates some of the more common analytical approximations to single-cell TCA models.Comment: Accepted for publication in "Physics Reports". A version of this paper with high-quality images can be found at: http://phdsven.dyns.cx (go to "Papers written"

    Higher Dimensional Energetic Causal Sets

    Full text link
    The energetic causal set (ECS) program of Cort\^es and Smolin, whose distinguishing feature is the foundational time irreversibility of the evolution equations, was initiated ten years ago. The toy model used to show the emergence of the time-reversible phase from the time-irreversible foundational regime originally had only one spatial dimension (1d+1). The extension to two and more spatial dimensions has posed a substantial challenge, where the higher-dimensional set of solutions had measure zero (involving infinitely-specified initial conditions). This challenge is overcome here with the extension of the ECS to 2+1 dimensions, invoking a finite interaction cross-section to determine generation of new events and including an adjustable fundamental stochasticity. Successfully, as in the 1+1 dimensional case, we observe a phase transition into a time-symmetric phase, here through the emergence of crystal-like structures. Due to the irreversible evolution we also witness the discrete dynamical systems (DDS) behaviour of the 1d+1 case explored in later articles of the ECS program, in which the model is captured in and out of the limit cycles which here take the form of lattice crystals. In a companion paper we carry out a detailed parameter investigation and study the causal network underlying the set, explaining the emergence of the time-symmetric phase.Comment: 15 pages, 19 figures, updated to improve figure quality and layou

    Fast in-memory XPath search using compressed indexes

    Get PDF
    A large fraction of an XML document typically consists of text data. The XPath query language allows text search via the equal, contains, and starts-with predicates. Such predicates can be efficiently implemented using a compressed self-index of the document's text nodes. Most queries, however, contain some parts querying the text of the document, plus some parts querying the tree structure. It is therefore a challenge to choose an appropriate evaluation order for a given query, which optimally leverages the execution speeds of the text and tree indexes. Here the SXSI system is introduced. It stores the tree structure of an XML document using a bit array of opening and closing brackets plus a sequence of labels, and stores the text nodes of the document using a global compressed self-index. On top of these indexes sits an XPath query engine that is based on tree automata. The engine uses fast counting queries of the text index in order to dynamically determine whether to evaluate top-down or bottom-up with respect to the tree structure. The resulting system has several advantages over existing systems: (1) on pure tree queries (without text search) such as the XPathMark queries, the SXSI system performs on par or better than the fastest known systems MonetDB and Qizx, (2) on queries that use text search, SXSI outperforms the existing systems by 1-3 orders of magnitude (depending on the size of the result set), and (3) with respect to memory consumption, SXSI outperforms all other systems for counting-only queries.Peer reviewe

    Selective Sampling with Information-Storage Constraints

    Get PDF
    A memoryless agent can acquire arbitrarily many signals. After each signal observation, she either terminates and chooses an action, or she discards her observation and draws a new signal. By conditioning the probability of termination on the information collected, she controls the correlation between the payoff state and her terminal action. We provide an optimality condition for the emerging stochastic choice. The condition highlights the benefits of selective memory applied to the extracted signals. Implications—obtained in simple examples—include (i) confirmation bias, (ii) speed-accuracy complementarity, (iii) overweighting of rare events, and (iv) salience effect

    Robustness and stability in dynamic constraint satisfaction problems

    Full text link
    Constraint programming is a paradigm wherein relations between variables are stated in the form of constraints. It is well-known that many real life problems can be modeled as Constraint Satisfaction Problems (CSPs). Much effort has been spent to increase the efficiency of algorithms for solving CSPs. However, many of these techniques assume that the set of variables, domains and constraints involved in the CSP are known and fixed when the problem is modeled. This is a strong limitation because many problems come from uncertain and dynamic environments, where both the original problem may evolve because of the environment, the user or other agents. In such situations, a solution that holds for the original problem can become invalid after changes. There are two main approaches for dealing with these situations: reactive and proactive approaches. Using reactive approaches entails re-solving the CSP after each solution loss, which is a time consuming. That is a clear disadvantage, especially when we deal with short-term changes, where solution loss is frequent. In addition, in many applications, such as on-line planning and scheduling, the delivery time of a new solution may be too long for actions to be taken on time, so a solution loss can produce several negative effects in the modeled problem. For a task assignment production system with several machines, it could cause the shutdown of the production system, the breakage of machines, the loss of the material/object in production, etc. In a transport timetabling problem, the solution loss, due to some disruption at a point, may produce a delay that propagates through the entire schedule. In addition, all the negative effects stated above will probably entail an economic loss. In this thesis we develop several proactive approaches. Proactive approaches use knowledge about possible future changes in order to avoid or minimize their effects. These approaches are applied before the changes occur. Thus, our approaches search for robust solutions, which have a high probability to remain valid after changes. Furthermore, some of our approaches also consider that the solutions can be easily adapted when they did not resist the changes in the original problem. Thus, these approaches search for stable solutions, which have an alternative solution that is similar to the previous one and therefore can be used in case of a value breakage. In this context, sometimes there exists knowledge about the uncertain and dynamic environment. However in many cases, this information is unknown or hard to obtain. For this reason, for the majority of our approaches (specifically 3 of the 4 developed approaches), the only assumptions made about changes are those inherent in the structure of problems with ordered domains. Given this framework and therefore the existence of a significant order over domain values, it is reasonable to assume that the original bounds of the solution space may undergo restrictive or relaxed modifications. Note that the possibility of solution loss only exists when changes over the original bounds of the solution space are restrictive. Therefore, the main objective for searching robust solutions in this framework is to find solutions located as far away as possible from the bounds of the solution space. In order to meet this criterion, we propose several approaches that can be divided in enumeration-based techniques and a search algorithm.Climent Aunés, LI. (2013). Robustness and stability in dynamic constraint satisfaction problems [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/34785TESI
    • …
    corecore