2,620 research outputs found
Mining Heterogeneous Multivariate Time-Series for Learning Meaningful Patterns: Application to Home Health Telecare
For the last years, time-series mining has become a challenging issue for
researchers. An important application lies in most monitoring purposes, which
require analyzing large sets of time-series for learning usual patterns. Any
deviation from this learned profile is then considered as an unexpected
situation. Moreover, complex applications may involve the temporal study of
several heterogeneous parameters. In that paper, we propose a method for mining
heterogeneous multivariate time-series for learning meaningful patterns. The
proposed approach allows for mixed time-series -- containing both pattern and
non-pattern data -- such as for imprecise matches, outliers, stretching and
global translating of patterns instances in time. We present the early results
of our approach in the context of monitoring the health status of a person at
home. The purpose is to build a behavioral profile of a person by analyzing the
time variations of several quantitative or qualitative parameters recorded
through a provision of sensors installed in the home
Composing Distributed Data-intensive Web Services Using a Flexible Memetic Algorithm
Web Service Composition (WSC) is a particularly promising application of Web
services, where multiple individual services with specific functionalities are
composed to accomplish a more complex task, which must fulfil functional
requirements and optimise Quality of Service (QoS) attributes, simultaneously.
Additionally, large quantities of data, produced by technological advances,
need to be exchanged between services. Data-intensive Web services, which
manipulate and deal with those data, are of great interest to implement
data-intensive processes, such as distributed Data-intensive Web Service
Composition (DWSC). Researchers have proposed Evolutionary Computing (EC)
fully-automated WSC techniques that meet all the above factors. Some of these
works employed Memetic Algorithms (MAs) to enhance the performance of EC
through increasing its exploitation ability of in searching neighbourhood area
of a solution. However, those works are not efficient or effective. This paper
proposes an MA-based approach to solving the problem of distributed DWSC in an
effective and efficient manner. In particular, we develop an MA that hybridises
EC with a flexible local search technique incorporating distance of services.
An evaluation using benchmark datasets is carried out, comparing existing
state-of-the-art methods. Results show that our proposed method has the highest
quality and an acceptable execution time overall.Comment: arXiv admin note: text overlap with arXiv:1901.0556
Longest Increasing Subsequence under Persistent Comparison Errors
We study the problem of computing a longest increasing subsequence in a
sequence of distinct elements in the presence of persistent comparison
errors. In this model, every comparison between two elements can return the
wrong result with some fixed (small) probability , and comparisons cannot
be repeated. Computing the longest increasing subsequence exactly is impossible
in this model, therefore, the objective is to identify a subsequence that (i)
is indeed increasing and (ii) has a length that approximates the length of the
longest increasing subsequence.
We present asymptotically tight upper and lower bounds on both the
approximation factor and the running time. In particular, we present an
algorithm that computes an -approximation in time , with
high probability. This approximation relies on the fact that that we can
approximately sort elements in time such that the maximum
dislocation of an element is at most . For the lower bounds, we
prove that (i) there is a set of sequences, such that on a sequence picked
randomly from this set every algorithm must return an -approximation with high probability, and (ii) any -approximation
algorithm for longest increasing subsequence requires
comparisons, even in the absence of errors
Run Generation Revisited: What Goes Up May or May Not Come Down
In this paper, we revisit the classic problem of run generation. Run
generation is the first phase of external-memory sorting, where the objective
is to scan through the data, reorder elements using a small buffer of size M ,
and output runs (contiguously sorted chunks of elements) that are as long as
possible.
We develop algorithms for minimizing the total number of runs (or
equivalently, maximizing the average run length) when the runs are allowed to
be sorted or reverse sorted. We study the problem in the online setting, both
with and without resource augmentation, and in the offline setting.
(1) We analyze alternating-up-down replacement selection (runs alternate
between sorted and reverse sorted), which was studied by Knuth as far back as
1963. We show that this simple policy is asymptotically optimal. Specifically,
we show that alternating-up-down replacement selection is 2-competitive and no
deterministic online algorithm can perform better.
(2) We give online algorithms having smaller competitive ratios with resource
augmentation. Specifically, we exhibit a deterministic algorithm that, when
given a buffer of size 4M , is able to match or beat any optimal algorithm
having a buffer of size M . Furthermore, we present a randomized online
algorithm which is 7/4-competitive when given a buffer twice that of the
optimal.
(3) We demonstrate that performance can also be improved with a small amount
of foresight. We give an algorithm, which is 3/2-competitive, with
foreknowledge of the next 3M elements of the input stream. For the extreme case
where all future elements are known, we design a PTAS for computing the optimal
strategy a run generation algorithm must follow.
(4) Finally, we present algorithms tailored for nearly sorted inputs which
are guaranteed to have optimal solutions with sufficiently long runs
Streaming and Query Once Space Complexity of Longest Increasing Subsequence
Longest Increasing Subsequence (LIS) is a fundamental problem in
combinatorics and computer science. Previously, there have been numerous works
on both upper bounds and lower bounds of the time complexity of computing and
approximating LIS, yet only a few on the equally important space complexity.
In this paper, we further study the space complexity of computing and
approximating LIS in various models. Specifically, we prove non-trivial space
lower bounds in the following two models: (1) the adaptive query-once model or
read-once branching programs, and (2) the streaming model where the order of
streaming is different from the natural order.
As far as we know, there are no previous works on the space complexity of LIS
in these models. Besides the bounds, our work also leaves many intriguing open
problems.Comment: This paper has been accepted to COCOON 202
- …