1,385 research outputs found

    Temporalized logics and automata for time granularity

    Full text link
    Suitable extensions of the monadic second-order theory of k successors have been proposed in the literature to capture the notion of time granularity. In this paper, we provide the monadic second-order theories of downward unbounded layered structures, which are infinitely refinable structures consisting of a coarsest domain and an infinite number of finer and finer domains, and of upward unbounded layered structures, which consist of a finest domain and an infinite number of coarser and coarser domains, with expressively complete and elementarily decidable temporal logic counterparts. We obtain such a result in two steps. First, we define a new class of combined automata, called temporalized automata, which can be proved to be the automata-theoretic counterpart of temporalized logics, and show that relevant properties, such as closure under Boolean operations, decidability, and expressive equivalence with respect to temporal logics, transfer from component automata to temporalized ones. Then, we exploit the correspondence between temporalized logics and automata to reduce the task of finding the temporal logic counterparts of the given theories of time granularity to the easier one of finding temporalized automata counterparts of them.Comment: Journal: Theory and Practice of Logic Programming Journal Acronym: TPLP Category: Paper for Special Issue (Verification and Computational Logic) Submitted: 18 March 2002, revised: 14 Januari 2003, accepted: 5 September 200

    Platform-level Distributed Warfare Model-based on Multi-Agent System Framework

    Get PDF
    The multi-agent paradigm has become a useful tool in solving military problems. However, one of key challenges in multi-agent model for distributed warfare could be how to describe the microcosmic  tactical warfare platforms actions. In this paper, a platform-level distributed warfare model based on multi-agent system framework is designed to tackle this challenge. The basic ideas include:  Establishing multi-agent model by mapping from tactical warfare system’s members, i.e., warfare platforms, to respective agents; performing task decomposition and task allocation by using task-tree decomposition method and improved contract net protocol model technique; and implementing simulation by presenting battlefield terrain environment analysis algorithm based on grid approach. The  simulation demonstration results show that our model provides a feasible and effective approach to supporting the abstraction and representation of microcosmic tactical actions for complex warfare system.Defence Science Journal, 2012, 62(1), pp.180-186, DOI:http://dx.doi.org/10.14429/dsj.62.96

    Multiscaled Cross-Correlation Dynamics in Financial Time-Series

    Get PDF
    The cross correlation matrix between equities comprises multiple interactions between traders with varying strategies and time horizons. In this paper, we use the Maximum Overlap Discrete Wavelet Transform to calculate correlation matrices over different timescales and then explore the eigenvalue spectrum over sliding time windows. The dynamics of the eigenvalue spectrum at different times and scales provides insight into the interactions between the numerous constituents involved. Eigenvalue dynamics are examined for both medium and high-frequency equity returns, with the associated correlation structure shown to be dependent on both time and scale. Additionally, the Epps effect is established using this multivariate method and analyzed at longer scales than previously studied. A partition of the eigenvalue time-series demonstrates, at very short scales, the emergence of negative returns when the largest eigenvalue is greatest. Finally, a portfolio optimization shows the importance of timescale information in the context of risk management

    Time granularity in simulation models within a multi-agent system

    Get PDF
    The understanding of how processes in natural phenomena interact at different scales of time has been a great challenge for humans. How information is transferred across scales is fundamental if one tries to scale up from finer to coarse levels of granularity. Computer simulation has been a powerful tool to determine the appropriate amount of detail one has to impose when developing simulation models of such phenomena. However, it has proved difficult to represent change at many scales of time and subject to cyclical processes. This issue has received little attention in traditional AI work on temporal reasoning but it becomes important in more complex domains, such as ecological modelling. Traditionally, models of ecosystems have been developed using imperative languages. Very few of those temporal logic theories have been used for the specification of simulation models in ecology. The aggregation of processes working at different scales of time is difficult (sometimes impossible) to do reliably. The reason is because these processes influence each other, and their functionality does not always scale to other levels. Thus the problems to tackle are representing cyclical and interacting processes at many scales and providing a framework to make the integration of such processes more reliable. We propose a framework for temporal modelling which allows modellers to represent cyclical and interacting processes at many scales. This theory combines both aspects by means of modular temporal classes and an underlying special temporal unification algorithm. To allow integration of different models they are developed as agents with a degree of autonomy in a multi-agent system architecture. This Ecoagency framework is evaluated on ecological modelling problems and it is compared to a formal language for describing ecological systems

    A Market-based Approach to Multi-factory Scheduling

    No full text
    In this paper, we report on the design of a novel market-based approach for decentralised scheduling across multiple factories. Specifically, because of the limitations of scheduling in a centralised manner -- which requires a center to have complete and perfect information for optimality and the truthful revelation of potentially commercially private preferences to that center -- we advocate an informationally decentralised approach that is both agile and dynamic. In particular, this work adopts a market-based approach for decentralised scheduling by considering the different stakeholders representing different factories as self-interested, profit-motivated economic agents that trade resources for the scheduling of jobs. The overall schedule of these jobs is then an emergent behaviour of the strategic interaction of these trading agents bidding for resources in a market based on limited information and their own preferences. Using a simple (zero-intelligence) bidding strategy, we empirically demonstrate that our market-based approach achieves a lower bound efficiency of 84%. This represents a trade-off between a reasonable level of efficiency (compared to a centralised approach) and the desirable benefits of a decentralised solution

    The Emergence of Norms via Contextual Agreements in Open Societies

    Full text link
    This paper explores the emergence of norms in agents' societies when agents play multiple -even incompatible- roles in their social contexts simultaneously, and have limited interaction ranges. Specifically, this article proposes two reinforcement learning methods for agents to compute agreements on strategies for using common resources to perform joint tasks. The computation of norms by considering agents' playing multiple roles in their social contexts has not been studied before. To make the problem even more realistic for open societies, we do not assume that agents share knowledge on their common resources. So, they have to compute semantic agreements towards performing their joint actions. %The paper reports on an empirical study of whether and how efficiently societies of agents converge to norms, exploring the proposed social learning processes w.r.t. different society sizes, and the ways agents are connected. The results reported are very encouraging, regarding the speed of the learning process as well as the convergence rate, even in quite complex settings

    Proceedings of the ECSCW'95 Workshop on the Role of Version Control in CSCW Applications

    Full text link
    The workshop entitled "The Role of Version Control in Computer Supported Cooperative Work Applications" was held on September 10, 1995 in Stockholm, Sweden in conjunction with the ECSCW'95 conference. Version control, the ability to manage relationships between successive instances of artifacts, organize those instances into meaningful structures, and support navigation and other operations on those structures, is an important problem in CSCW applications. It has long been recognized as a critical issue for inherently cooperative tasks such as software engineering, technical documentation, and authoring. The primary challenge for versioning in these areas is to support opportunistic, open-ended design processes requiring the preservation of historical perspectives in the design process, the reuse of previous designs, and the exploitation of alternative designs. The primary goal of this workshop was to bring together a diverse group of individuals interested in examining the role of versioning in Computer Supported Cooperative Work. Participation was encouraged from members of the research community currently investigating the versioning process in CSCW as well as application designers and developers who are familiar with the real-world requirements for versioning in CSCW. Both groups were represented at the workshop resulting in an exchange of ideas and information that helped to familiarize developers with the most recent research results in the area, and to provide researchers with an updated view of the needs and challenges faced by application developers. In preparing for this workshop, the organizers were able to build upon the results of their previous one entitled "The Workshop on Versioning in Hypertext" held in conjunction with the ECHT'94 conference. The following section of this report contains a summary in which the workshop organizers report the major results of the workshop. The summary is followed by a section that contains the position papers that were accepted to the workshop. The position papers provide more detailed information describing recent research efforts of the workshop participants as well as current challenges that are being encountered in the development of CSCW applications. A list of workshop participants is provided at the end of the report. The organizers would like to thank all of the participants for their contributions which were, of course, vital to the success of the workshop. We would also like to thank the ECSCW'95 conference organizers for providing a forum in which this workshop was possible

    Ontology-Based Consistent Specification of Sensor Data Acquisition Plans in Cross-Domain IoT Platforms

    Get PDF
    Nowadays there is an high number of IoT applications that seldom can interact with each other because developed within different Vertical IoT Platforms that adopt different standards. Several efforts are devoted to the construction of cross-layered frameworks that facilitate the interoperability among cross-domain IoT platforms for the development of horizontal applications. Even if their realization poses different challenges across all layers of the network stack, in this paper we focus on the interoperability issues that arise at the data management layer. Specifically, starting from a flexible multi-granular Spatio-Temporal-Thematic data model according to which events generated by different kinds of sensors can be represented, we propose a Semantic Virtualization approach according to which the sensors belonging to different IoT platforms and the schema of the produced event streams are described in a Domain Ontology, obtained through the extension of the well-known Semantic Sensor Network ontology. Then, these sensors can be exploited for the creation of Data Acquisition Plans by means of which the streams of events can be filtered, merged, and aggregated in a meaningful way. A notion of consistency is introduced to bind the output streams of the services contained in the Data Acquisition Plan with the Domain Ontology in order to provide a semantic description of its final output. When these plans meet the consistency constraints, it means that the data they handle are well described at the Ontological level and thus the data acquisition process over passed the interoperability barriers occurring in the original sources. The facilities of the StreamLoader prototype are finally presented for supporting the user in the Semantic Virtualization process and for the construction of meaningful Data Acquisition Plans
    corecore