3,883 research outputs found
Transforming Energy Networks via Peer to Peer Energy Trading: Potential of Game Theoretic Approaches
Peer-to-peer (P2P) energy trading has emerged as a next-generation energy
management mechanism for the smart grid that enables each prosumer of the
network to participate in energy trading with one another and the grid. This
poses a significant challenge in terms of modeling the decision-making process
of each participant with conflicting interest and motivating prosumers to
participate in energy trading and to cooperate, if necessary, for achieving
different energy management goals. Therefore, such decision-making process
needs to be built on solid mathematical and signal processing tools that can
ensure an efficient operation of the smart grid. This paper provides an
overview of the use of game theoretic approaches for P2P energy trading as a
feasible and effective means of energy management. As such, we discuss various
games and auction theoretic approaches by following a systematic classification
to provide information on the importance of game theory for smart energy
research. Then, the paper focuses on the P2P energy trading describing its key
features and giving an introduction to an existing P2P testbed. Further, the
paper zooms into the detail of some specific game and auction theoretic models
that have recently been used in P2P energy trading and discusses some important
finding of these schemes.Comment: 38 pages, single column, double spac
An Information-Theoretic Framework for Consistency Maintenance in Distributed Interactive Applications
Distributed Interactive Applications (DIAs) enable geographically dispersed users
to interact with each other in a virtual environment. A key factor to the success
of a DIA is the maintenance of a consistent view of the shared virtual world for
all the participants. However, maintaining consistent states in DIAs is difficult
under real networks. State changes communicated by messages over such networks
suffer latency leading to inconsistency across the application. Predictive Contract
Mechanisms (PCMs) combat this problem through reducing the number of messages
transmitted in return for perceptually tolerable inconsistency. This thesis examines
the operation of PCMs using concepts and methods derived from information theory.
This information theory perspective results in a novel information model of PCMs
that quantifies and analyzes the efficiency of such methods in communicating the
reduced state information, and a new adaptive multiple-model-based framework for
improving consistency in DIAs.
The first part of this thesis introduces information measurements of user behavior
in DIAs and formalizes the information model for PCM operation. In presenting the
information model, the statistical dependence in the entity state, which makes using
extrapolation models to predict future user behavior possible, is evaluated. The
efficiency of a PCM to exploit such predictability to reduce the amount of network
resources required to maintain consistency is also investigated. It is demonstrated
that from the information theory perspective, PCMs can be interpreted as a form
of information reduction and compression.
The second part of this thesis proposes an Information-Based Dynamic Extrapolation
Model for dynamically selecting between extrapolation algorithms based on
information evaluation and inferred network conditions. This model adapts PCM
configurations to both user behavior and network conditions, and makes the most
information-efficient use of the available network resources. In doing so, it improves
PCM performance and consistency in DIAs
Discrete event simulation and virtual reality use in industry: new opportunities and future trends
This paper reviews the area of combined discrete
event simulation (DES) and virtual reality (VR) use within industry.
While establishing a state of the art for progress in this
area, this paper makes the case for VR DES as the vehicle of choice
for complex data analysis through interactive simulation models,
highlighting both its advantages and current limitations. This paper
reviews active research topics such as VR and DES real-time
integration, communication protocols, system design considerations,
model validation, and applications of VR and DES. While
summarizing future research directions for this technology combination,
the case is made for smart factory adoption of VR DES as
a new platform for scenario testing and decision making. It is put
that in order for VR DES to fully meet the visualization requirements
of both Industry 4.0 and Industrial Internet visions of digital
manufacturing, further research is required in the areas of lower
latency image processing, DES delivery as a service, gesture recognition
for VR DES interaction, and linkage of DES to real-time data streams and Big Data sets
Proceedings of International Workshop "Global Computing: Programming Environments, Languages, Security and Analysis of Systems"
According to the IST/ FET proactive initiative on GLOBAL COMPUTING, the goal is to obtain techniques (models, frameworks, methods, algorithms) for constructing systems that are flexible, dependable, secure, robust and efficient.
The dominant concerns are not those of representing and manipulating data efficiently but rather those of handling the co-ordination and interaction, security, reliability, robustness, failure modes, and control of risk of the entities in the system and the overall design, description and performance of the system itself.
Completely different paradigms of computer science may have to be developed to tackle these issues effectively. The research should concentrate on systems having the following characteristics: • The systems are composed of autonomous computational entities where activity is not centrally controlled, either because global control is impossible or impractical, or because the entities are created or controlled by different owners.
• The computational entities are mobile, due to the movement of the physical platforms or by movement of the entity from one platform to another.
• The configuration varies over time. For instance, the system is open to the introduction of new computational entities and likewise their deletion.
The behaviour of the entities may vary over time.
• The systems operate with incomplete information about the environment.
For instance, information becomes rapidly out of date and mobility requires information about the environment to be discovered.
The ultimate goal of the research action is to provide a solid scientific foundation for the design of such systems, and to lay the groundwork for achieving effective principles for building and analysing such systems.
This workshop covers the aspects related to languages and programming environments as well as analysis of systems and resources involving 9 projects (AGILE , DART, DEGAS , MIKADO, MRG, MYTHS, PEPITO, PROFUNDIS, SECURE) out of the 13 founded under the initiative. After an year from the start of the projects, the goal of the workshop is to fix the state of the art on the topics covered by the two clusters related to programming environments and analysis of systems as well as to devise strategies and new ideas to profitably continue the research effort towards the overall objective of the initiative.
We acknowledge the Dipartimento di Informatica and Tlc of the University of Trento, the Comune di Rovereto, the project DEGAS for partially funding the event and the Events and Meetings Office of the University of Trento for the valuable collaboration
Safe, Remote-Access Swarm Robotics Research on the Robotarium
This paper describes the development of the Robotarium -- a remotely
accessible, multi-robot research facility. The impetus behind the Robotarium is
that multi-robot testbeds constitute an integral and essential part of the
multi-agent research cycle, yet they are expensive, complex, and time-consuming
to develop, operate, and maintain. These resource constraints, in turn, limit
access for large groups of researchers and students, which is what the
Robotarium is remedying by providing users with remote access to a
state-of-the-art multi-robot test facility. This paper details the design and
operation of the Robotarium as well as connects these to the particular
considerations one must take when making complex hardware remotely accessible.
In particular, safety must be built in already at the design phase without
overly constraining which coordinated control programs the users can upload and
execute, which calls for minimally invasive safety routines with provable
performance guarantees.Comment: 13 pages, 7 figures, 3 code samples, 72 reference
Information fusion architectures for security and resource management in cyber physical systems
Data acquisition through sensors is very crucial in determining the operability of the observed physical entity. Cyber Physical Systems (CPSs) are an example of distributed systems where sensors embedded into the physical system are used in sensing and data acquisition. CPSs are a collaboration between the physical and the computational cyber components. The control decisions sent back to the actuators on the physical components from the computational cyber components closes the feedback loop of the CPS. Since, this feedback is solely based on the data collected through the embedded sensors, information acquisition from the data plays an extremely vital role in determining the operational stability of the CPS. Data collection process may be hindered by disturbances such as system faults, noise and security attacks. Hence, simple data acquisition techniques will not suffice as accurate system representation cannot be obtained. Therefore, more powerful methods of inferring information from collected data such as Information Fusion have to be used.
Information fusion is analogous to the cognitive process used by humans to integrate data continuously from their senses to make inferences about their environment. Data from the sensors is combined using techniques drawn from several disciplines such as Adaptive Filtering, Machine Learning and Pattern Recognition. Decisions made from such combination of data form the crux of information fusion and differentiates it from a flat structured data aggregation. In this dissertation, multi-layered information fusion models are used to develop automated decision making architectures to service security and resource management requirements in Cyber Physical Systems --Abstract, page iv
A Game Theoretic approach based virtual machine migration for cloud environment security
In cloud computing environment, static configurations can provide for the attackers an environment too easy for exploitation and discovering the network vulnerabilities in order to compromise the network and launching intrusions; while dynamic reconfiguration seeks to develop a virtual machine (VM) migration over the cloud by applying unpredictability of network configuration’s change, and thus improving the system security. In this work a novel approach that performs proactive and reactive measures to ensure a high availability and to minimize the attack surface using VM migration is proposed. This interaction between attack and defense systems was formulated as game model. As result, we have calculated the Nash equilibrium and the utilities for the both attacker and defender, evaluate the parameters which can maximize the defender’s utility when the VM migration was planned and identify the potential attack paths. Therefore, the effectiveness of the game model was validated by some numerical results that determine optimal migration strategies in order to ensure the security of the system
- …