14,483 research outputs found
D-SLATS: Distributed Simultaneous Localization and Time Synchronization
Through the last decade, we have witnessed a surge of Internet of Things
(IoT) devices, and with that a greater need to choreograph their actions across
both time and space. Although these two problems, namely time synchronization
and localization, share many aspects in common, they are traditionally treated
separately or combined on centralized approaches that results in an ineffcient
use of resources, or in solutions that are not scalable in terms of the number
of IoT devices. Therefore, we propose D-SLATS, a framework comprised of three
different and independent algorithms to jointly solve time synchronization and
localization problems in a distributed fashion. The First two algorithms are
based mainly on the distributed Extended Kalman Filter (EKF) whereas the third
one uses optimization techniques. No fusion center is required, and the devices
only communicate with their neighbors. The proposed methods are evaluated on
custom Ultra-Wideband communication Testbed and a quadrotor, representing a
network of both static and mobile nodes. Our algorithms achieve up to three
microseconds time synchronization accuracy and 30 cm localization error
Parallel Deferred Update Replication
Deferred update replication (DUR) is an established approach to implementing
highly efficient and available storage. While the throughput of read-only
transactions scales linearly with the number of deployed replicas in DUR, the
throughput of update transactions experiences limited improvements as replicas
are added. This paper presents Parallel Deferred Update Replication (P-DUR), a
variation of classical DUR that scales both read-only and update transactions
with the number of cores available in a replica. In addition to introducing the
new approach, we describe its full implementation and compare its performance
to classical DUR and to Berkeley DB, a well-known standalone database
A Stochastic Model of Plausibility in Live-Virtual-Constructive Environments
Distributed live-virtual-constructive simulation promises a number of benefits for the test and evaluation community, including reduced costs, access to simulations of limited availability assets, the ability to conduct large-scale multi-service test events, and recapitalization of existing simulation investments. However, geographically distributed systems are subject to fundamental state consistency limitations that make assessing the data quality of live-virtual-constructive experiments difficult. This research presents a data quality model based on the notion of plausible interaction outcomes. This model explicitly accounts for the lack of absolute state consistency in distributed real-time systems and offers system designers a means of estimating data quality and fitness for purpose. Experiments with World of Warcraft player trace data validate the plausibility model and exceedance probability estimates. Additional experiments with synthetic data illustrate the model\u27s use in ensuring fitness for purpose of live-virtual-constructive simulations and estimating the quality of data obtained from live-virtual-constructive experiments
Middleware-based Database Replication: The Gaps between Theory and Practice
The need for high availability and performance in data management systems has
been fueling a long running interest in database replication from both academia
and industry. However, academic groups often attack replication problems in
isolation, overlooking the need for completeness in their solutions, while
commercial teams take a holistic approach that often misses opportunities for
fundamental innovation. This has created over time a gap between academic
research and industrial practice.
This paper aims to characterize the gap along three axes: performance,
availability, and administration. We build on our own experience developing and
deploying replication systems in commercial and academic settings, as well as
on a large body of prior related work. We sift through representative examples
from the last decade of open-source, academic, and commercial database
replication systems and combine this material with case studies from real
systems deployed at Fortune 500 customers. We propose two agendas, one for
academic research and one for industrial R&D, which we believe can bridge the
gap within 5-10 years. This way, we hope to both motivate and help researchers
in making the theory and practice of middleware-based database replication more
relevant to each other.Comment: 14 pages. Appears in Proc. ACM SIGMOD International Conference on
Management of Data, Vancouver, Canada, June 200
- …