8,039 research outputs found
Query Results over Ongoing Databases that Remain Valid as Time Passes By (Extended Version)
Ongoing time point now is used to state that a tuple is valid from the start
point onward. For database systems ongoing time points have far-reaching
implications since they change continuously as time passes by. State-of-the-art
approaches deal with ongoing time points by instantiating them to the reference
time. The instantiation yields query results that are only valid at the chosen
time and get invalidated as time passes by. We propose a solution that keeps
ongoing time points uninstantiated during query processing. We do so by
evaluating predicates and functions at all possible reference times. This
renders query results independent of a specific reference time and yields
results that remain valid as time passes by. As query results, we propose
ongoing relations that include a reference time attribute. The value of the
reference time attribute is restricted by predicates and functions on ongoing
attributes. We describe and evaluate an efficient implementation of ongoing
data types and operations in PostgreSQL.Comment: Extended version of ICDE pape
The Family of MapReduce and Large Scale Data Processing Systems
In the last two decades, the continuous increase of computational power has
produced an overwhelming flow of data which has called for a paradigm shift in
the computing architecture and large scale data processing mechanisms.
MapReduce is a simple and powerful programming model that enables easy
development of scalable parallel applications to process vast amounts of data
on large clusters of commodity machines. It isolates the application from the
details of running a distributed program such as issues on data distribution,
scheduling and fault tolerance. However, the original implementation of the
MapReduce framework had some limitations that have been tackled by many
research efforts in several followup works after its introduction. This article
provides a comprehensive survey for a family of approaches and mechanisms of
large scale data processing mechanisms that have been implemented based on the
original idea of the MapReduce framework and are currently gaining a lot of
momentum in both research and industrial communities. We also cover a set of
introduced systems that have been implemented to provide declarative
programming interfaces on top of the MapReduce framework. In addition, we
review several large scale data processing systems that resemble some of the
ideas of the MapReduce framework for different purposes and application
scenarios. Finally, we discuss some of the future research directions for
implementing the next generation of MapReduce-like solutions.Comment: arXiv admin note: text overlap with arXiv:1105.4252 by other author
A framework for integrating DNA sequenced data
The Human Genome Project generated vast amounts of DNA sequenced data scattered in disparate data sources in a variety of formats. Integrating biological data and extracting information held in DNA sequences are major ongoing tasks for biologists and software professionals. This thesis explored issues of finding, extracting, merging and synthesizing information from multiple disparate data sources containing DNA sequenced data, which is composed of 3 billion chemical building blocks of bases. We proposed a biological data integration framework based on typical usage patterns to simplify these issues for biologists. The framework uses a relational database management system at the backend, and provides techniques to extract, store, and manage the data. This framework was implemented, evaluated, and compared with existing biological data integration solutions
Develop a generic Rules Engine to quality control a CV database
This bachelor’s thesis presents a software solution to enhance Bouvet’s quality control process
for employee CVs. By implementing a generic rule engine with extended functionalities, we
identified that 90% of the CVs at Bouvet did not meet the company’s business standards.
Using Scrum with Extreme Programming as our project management system, we developed a
scalable and maintainable pilot, employing Microservices, Event-Driven, and Command and
Query Responsibility Segregation architecture. Our pilot allows for future modifications using
create, read, update and delete operations. The software solution presented in this thesis can
be extended to a production-ready state by implementing an Role-based access control and
an API-Gateway. When the event bus project by another group at Bouvet is completed, our
implementation will be able to notify employees about their CVs’ status, further improving
the quality control process. Overall, our results demonstrate the our software solution and
project management system in enhancing the quality control of employee CVs at Bouvet.This bachelor’s thesis presents a software solution to enhance Bouvet’s quality control process
for employee CVs. By implementing a generic rule engine with extended functionalities, we
identified that 90% of the CVs at Bouvet did not meet the company’s business standards.
Using Scrum with Extreme Programming as our project management system, we developed a
scalable and maintainable pilot, employing Microservices, Event-Driven, and Command and
Query Responsibility Segregation architecture. Our pilot allows for future modifications using
create, read, update and delete operations. The software solution presented in this thesis can
be extended to a production-ready state by implementing an Role-based access control and
an API-Gateway. When the event bus project by another group at Bouvet is completed, our
implementation will be able to notify employees about their CVs’ status, further improving
the quality control process. Overall, our results demonstrate the our software solution and
project management system in enhancing the quality control of employee CVs at Bouvet
Develop a generic Rules Engine to quality control a CV database
This bachelor’s thesis presents a software solution to enhance Bouvet’s quality control process
for employee CVs. By implementing a generic rule engine with extended functionalities, we
identified that 90% of the CVs at Bouvet did not meet the company’s business standards.
Using Scrum with Extreme Programming as our project management system, we developed a
scalable and maintainable pilot, employing Microservices, Event-Driven, and Command and
Query Responsibility Segregation architecture. Our pilot allows for future modifications using
create, read, update and delete operations. The software solution presented in this thesis can
be extended to a production-ready state by implementing an Role-based access control and
an API-Gateway. When the event bus project by another group at Bouvet is completed, our
implementation will be able to notify employees about their CVs’ status, further improving
the quality control process. Overall, our results demonstrate the our software solution and
project management system in enhancing the quality control of employee CVs at Bouvet.This bachelor’s thesis presents a software solution to enhance Bouvet’s quality control process
for employee CVs. By implementing a generic rule engine with extended functionalities, we
identified that 90% of the CVs at Bouvet did not meet the company’s business standards.
Using Scrum with Extreme Programming as our project management system, we developed a
scalable and maintainable pilot, employing Microservices, Event-Driven, and Command and
Query Responsibility Segregation architecture. Our pilot allows for future modifications using
create, read, update and delete operations. The software solution presented in this thesis can
be extended to a production-ready state by implementing an Role-based access control and
an API-Gateway. When the event bus project by another group at Bouvet is completed, our
implementation will be able to notify employees about their CVs’ status, further improving
the quality control process. Overall, our results demonstrate the our software solution and
project management system in enhancing the quality control of employee CVs at Bouvet
Architecture for Provenance Systems
This document covers the logical and process architectures of provenance systems. The logical architecture identifies key roles and their interactions, whereas the process architecture discusses distribution and security. A fundamental aspect of our presentation is its technology-independent nature, which makes it reusable: the principles that are exposed in this document may be applied to different technologies
Ethical issues of electronic patient data and informatics in clinical trial settings
The field of cancer bio-informatics unites the disciplines of scientific and clinical research withclinical practice and the treatment of individual patients. There is a need to study patients andsometimes their families, over many decades, to follow disease progress and long-term outcomes.This may require research teams to access the routinely-collected health data from generalpractice and hospital health records, prior to and after the cancer diagnosis is made. This clinicalinformation will increasingly include data provided by patients or acquired from them throughwearable devices that can monitor or deliver treatment, and data acquired from genetic relativesof the patient.All of these data, whether explicitly collected for the purpose of a clinical study, or routinelycollected as part of a patient?s life-time healthcare journey, are personal health data. There areethical and legal requirements to manage these data with care. This chapter explores the ethicalrequirements for collecting, holding, analysing and sharing personal health data, and thelegislation covering such activities
Moving Target Defense for Web Applications
abstract: Web applications continue to remain as the most popular method of interaction for businesses over the Internet. With it's simplicity of use and management, they often function as the "front door" for many companies. As such, they are a critical component of the security ecosystem as vulnerabilities present in these systems could potentially allow malicious users access to sensitive business and personal data.
The inherent nature of web applications enables anyone to access them anytime and anywhere, this includes any malicious actors looking to exploit vulnerabilities present in the web application. In addition, the static configurations of these web applications enables attackers the opportunity to perform reconnaissance at their leisure, increasing their success rate by allowing them time to discover information on the system. On the other hand, defenders are often at a disadvantage as they do not have the same temporal opportunity that attackers possess in order to perform counter-reconnaissance. Lastly, the unchanging nature of web applications results in undiscovered vulnerabilities to remain open for exploitation, requiring developers to adopt a reactive approach that is often delayed or to anticipate and prepare for all possible attacks which is often cost-prohibitive.
Moving Target Defense (MTD) seeks to remove the attackers' advantage by reducing the information asymmetry between the attacker and defender. This research explores the concept of MTD and the various methods of applying MTD to secure Web Applications. In particular, MTD concepts are applied to web applications by implementing an automated application diversifier that aims to mitigate specific classes of web application vulnerabilities and exploits. Evaluation is done using two open source web applications to determine the effectiveness of the MTD implementation. Though developed for the chosen applications, the automation process can be customized to fit a variety of applications.Dissertation/ThesisMasters Thesis Computer Science 201
The MGX framework for microbial community analysis
Jaenicke S. The MGX framework for microbial community analysis. Bielefeld: Universität Bielefeld; 2020
- …