36 research outputs found
Worst-case response time of aperiodc message in WorldFIP and its improvement in real-time capability
MCFlow: Middleware for Mixed-Criticality Distributed Real-Time Systems
Traditional fixed-priority scheduling analysis for periodic/sporadic task sets is based on the assumption that all tasks are equally critical to the correct operation of the system. Therefore, every task has to be schedulable under the scheduling policy, and estimates of tasks\u27 worst case execution times must be conservative in case a task runs longer than is usual. To address the significant under-utilization of a system\u27s resources under normal operating conditions that can arise from these assumptions, several \emph{mixed-criticality scheduling} approaches have been proposed. However, to date there has been no quantitative comparison of system schedulability or run-time overhead for the different approaches. In this dissertation, we present what is to our knowledge the first side-by-side implementation and evaluation of those approaches, for periodic and sporadic mixed-criticality tasks on uniprocessor or distributed systems, under a mixed-criticality scheduling model that is common to all these approaches. To make a fair evaluation of mixed-criticality scheduling, we also address some previously open issues and propose modifications to improve schedulability and correctness of particular approaches. To facilitate the development and evaluation of mixed-criticality applications, we have designed and developed a distributed real-time middleware, called MCFlow, for mixed-criticality end-to-end tasks running on multi-core platforms. The research presented in this dissertation provides the following contributions to the state of the art in real-time middleware: (1) an efficient component model through which dependent subtask graphs can be configured flexibly for execution within a single core, across cores of a common host, or spanning multiple hosts; (2) support for optimizations to inter-component communication to reduce data copying without sacrificing the ability to execute subtasks in parallel; (3) a strict separation of timing and functional concerns so that they can be configured independently; (4) an event dispatching architecture that uses lock free algorithms where possible to reduce memory contention, CPU context switching, and priority inversion; and (5) empirical evaluations of MCFlow itself and of different mixed criticality scheduling approaches both with a single host and end-to-end across multiple hosts. The results of our evaluation show that in terms of basic distributed real-time behavior MCFlow performs comparably to the state of the art TAO real-time object request broker when only one core is used and outperforms TAO when multiple cores are involved. We also identify and categorize different use cases under which different mixed criticality scheduling approaches are preferable
Performance assessment of real-time data management on wireless sensor networks
Technological advances in recent years have allowed the maturity of Wireless Sensor Networks
(WSNs), which aim at performing environmental monitoring and data collection. This sort of
network is composed of hundreds, thousands or probably even millions of tiny smart computers
known as wireless sensor nodes, which may be battery powered, equipped with sensors, a radio
transceiver, a Central Processing Unit (CPU) and some memory. However due to the small size and
the requirements of low-cost nodes, these sensor node resources such as processing power, storage
and especially energy are very limited.
Once the sensors perform their measurements from the environment, the problem of data
storing and querying arises. In fact, the sensors have restricted storage capacity and the on-going
interaction between sensors and environment results huge amounts of data. Techniques for data
storage and query in WSN can be based on either external storage or local storage. The external
storage, called warehousing approach, is a centralized system on which the data gathered by the
sensors are periodically sent to a central database server where user queries are processed. The
local storage, in the other hand called distributed approach, exploits the capabilities of sensors
calculation and the sensors act as local databases. The data is stored in a central database server
and in the devices themselves, enabling one to query both.
The WSNs are used in a wide variety of applications, which may perform certain operations on
collected sensor data. However, for certain applications, such as real-time applications, the sensor
data must closely reflect the current state of the targeted environment. However, the environment
changes constantly and the data is collected in discreet moments of time. As such, the collected
data has a temporal validity, and as time advances, it becomes less accurate, until it does not
reflect the state of the environment any longer. Thus, these applications must query and analyze
the data in a bounded time in order to make decisions and to react efficiently, such as industrial
automation, aviation, sensors network, and so on. In this context, the design of efficient real-time
data management solutions is necessary to deal with both time constraints and energy consumption.
This thesis studies the real-time data management techniques for WSNs. It particularly it focuses
on the study of the challenges in handling real-time data storage and query for WSNs and on the
efficient real-time data management solutions for WSNs.
First, the main specifications of real-time data management are identified and the available
real-time data management solutions for WSNs in the literature are presented. Secondly, in order to
provide an energy-efficient real-time data management solution, the techniques used to manage
data and queries in WSNs based on the distributed paradigm are deeply studied. In fact, many
research works argue that the distributed approach is the most energy-efficient way of managing
data and queries in WSNs, instead of performing the warehousing. In addition, this approach can provide quasi real-time query processing because the most current data will be retrieved from the
network.
Thirdly, based on these two studies and considering the complexity of developing, testing, and
debugging this kind of complex system, a model for a simulation framework of the real-time
databases management on WSN that uses a distributed approach and its implementation are
proposed. This will help to explore various solutions of real-time database techniques on WSNs
before deployment for economizing money and time. Moreover, one may improve the proposed
model by adding the simulation of protocols or place part of this simulator on another available
simulator. For validating the model, a case study considering real-time constraints as well as energy
constraints is discussed.
Fourth, a new architecture that combines statistical modeling techniques with the distributed
approach and a query processing algorithm to optimize the real-time user query processing are
proposed. This combination allows performing a query processing algorithm based on admission
control that uses the error tolerance and the probabilistic confidence interval as admission
parameters. The experiments based on real world data sets as well as synthetic data sets
demonstrate that the proposed solution optimizes the real-time query processing to save more
energy while meeting low latency.Fundação para a Ciência e Tecnologi
Novel techniques for location-cloaked applications
Location cloaking has been shown to be cost-effective in mitigating location privacy and safety risks. This strategy, however, has significant impact on the applications that rely on location information. They may suffer efficiency loss; some may not even work with reduced location resolution. This research investigates two problems. 1) How to process location-cloaked queries. Processing such queries incurs significant more workload for both server and client. While the server needs to retrieve more query results and transmit them to the client, the client downloading these results wastes its battery power because most of them are useless. To address these problems, we propose a suite of novel techniques including query decomposition, scheduling, and personalized air indexing. These techniques are integrated into a single unified platform that is capable of handling various types of queries. 2) How a node V can verify whether or not another node P indeed locates in a cloaking region it claims. This problem is challenging due to the fact that the process of location verification may allow V to refine P\u27s location within the region. We identify two types of attacks, transmission coverage attack and distance bounding attack. In the former, V refines a cloaking region by adjusting its transmission range to partially overlap with the region, whereas in the latter, by measuring the round trip time of its communication with P. We present two corresponding counter strategies, and built on top of them, propose a novel technique that allows P to participate in location verification while providing a certain level of guarantee that its cloaking region will not be refined during the process
Hibi no seikatsu ni tokekomi seikatsu no shitsu o kojo saseru sofutowea no tame no furemuwaku
制度:新 ; 報告番号:甲3287号 ; 学位の種類:博士(工学) ; 授与年月日:2011/2/25 ; 早大学位記番号:新559
Using embedded hardware monitor cores in critical computer systems
The integration of FPGA devices in many different architectures and services
makes monitoring and real time detection of errors an important concern in FPGA
system design. A monitor is a tool, or a set of tools, that facilitate analytic
measurements in observing a given system. The goal of these observations is
usually the performance analysis and optimisation, or the surveillance of the system.
However, System-on-Chip (SoC) based designs leave few points to attach external
tools such as logic analysers. Thus, an embedded error detection core that allows
observation of critical system nodes (such as processor cores and buses) should
enforce the operation of the FPGA-based system, in order to prevent system
failures. The core should not interfere with system performance and must ensure
timely detection of errors.
This thesis is an investigation onto how a robust hardware-monitoring module
can be efficiently integrated in a target PCI board (with FPGA-based application processing
features) which is part of a critical computing system. [Continues.
The Gamification of Crowdsourcing Systems: Empirical Investigations and Design
Recent developments in modern information and communication technologies have spawned two rising phenomena, gamification and crowdsourcing, which are increasingly being combined into gamified crowdsourcing systems. While a growing number of organizations employ crowdsourcing as a way to outsource tasks related to the inventing, producing, funding, or distributing of their products and services to the crowd – a large group of people reachable via the internet – crowdsourcing initiatives become enriched with design features from games to motivate the crowd to participate in these efforts. From a practical perspective, this combination seems intuitively appealing, since using gamification in crowdsourcing systems promises to increase motivations, participation and output quality, as well as to replace traditionally used financial incentives. However, people in large groups all have individual interests and motivations, which makes it complex to design gamification approaches for crowds. Further, crowdsourcing systems exist in various forms and are used for various tasks and problems, thus requiring different incentive mechanisms for different crowdsourcing types. The lack of a coherent understanding of the different facets of gamified crowdsourcing systems and the lack of knowledge about the motivational and behavioral effects of applying various types of gamification features in different crowdsourcing systems inhibit us from designing solutions that harness gamification’s full potential. Further, previous research canonically uses competitive gamification, although crowdsourcing systems often strive to produce cooperative outcomes. However, the potentially relevant field of cooperative gamification has to date barely been explored. With a specific focus on these shortcomings, this dissertation presents several studies to advance the understanding of using gamification in crowdsourcing systems