1,657 research outputs found
Middleware-based Database Replication: The Gaps between Theory and Practice
The need for high availability and performance in data management systems has
been fueling a long running interest in database replication from both academia
and industry. However, academic groups often attack replication problems in
isolation, overlooking the need for completeness in their solutions, while
commercial teams take a holistic approach that often misses opportunities for
fundamental innovation. This has created over time a gap between academic
research and industrial practice.
This paper aims to characterize the gap along three axes: performance,
availability, and administration. We build on our own experience developing and
deploying replication systems in commercial and academic settings, as well as
on a large body of prior related work. We sift through representative examples
from the last decade of open-source, academic, and commercial database
replication systems and combine this material with case studies from real
systems deployed at Fortune 500 customers. We propose two agendas, one for
academic research and one for industrial R&D, which we believe can bridge the
gap within 5-10 years. This way, we hope to both motivate and help researchers
in making the theory and practice of middleware-based database replication more
relevant to each other.Comment: 14 pages. Appears in Proc. ACM SIGMOD International Conference on
Management of Data, Vancouver, Canada, June 200
Jet Momentum Resolution for the CMS Experiment and Distributed Data Caching Strategies
Accurately measured jets are mandatory for precision measurements of the Standard Model of particle physics as well as for searches for new physics.
The increased instantaneous luminosity and center-of-mass energy at LHC Run 2 pose challenges for pileup mitigation and the measurement of jet characteristics.
This thesis concentrates on using Z + jets events to calibrate the energy scale of jets recorded by the CMS detector in 2018.
Furthermore, it proposes a new procedure for determining the jet momentum resolution using Z + jets events.
This procedure is expected to allow cross-checking complementary measurement approaches and increasing the accuracy of the jet momentum resolution at the CMS experiment.
Data-intensive end-user analyses in High Energy Physics such as the presented calibration of jets put enormous challenges on the computing infrastructure since requiring high data throughput.
Besides the particle physics analysis, this thesis also focuses on accelerating data processing within a distributed computing infrastructure via a coordinated distributed caching approach.
Coordinated placement of critical data within distributed caches and matching workflows to the most suitable host in terms of cached data allows for optimizing processing efficiency.
Improving the processing of data-intensive workflows aims at shortening turnaround cycles and thus deriving physics results, e.g. the jet calibration results, faster
Performance analysis of a database caching system in a grid environment
Tese de mestrado. Engenharia Informática. Faculdade de Engenharia. Universidade do Porto. 200
Helmholtz Portfolio Theme Large-Scale Data Management and Analysis (LSDMA)
The Helmholtz Association funded the "Large-Scale Data Management and Analysis" portfolio theme from 2012-2016. Four Helmholtz centres, six universities and another research institution in Germany joined to enable data-intensive science by optimising data life cycles in selected scientific communities. In our Data Life cycle Labs, data experts performed joint R&D together with scientific communities. The Data Services Integration Team focused on generic solutions applied by several communities
- …