97,370 research outputs found
Overlapping and Robust Edge-Colored Clustering in Hypergraphs
A recent trend in data mining has explored (hyper)graph clustering algorithms
for data with categorical relationship types. Such algorithms have applications
in the analysis of social, co-authorship, and protein interaction networks, to
name a few. Many such applications naturally have some overlap between
clusters, a nuance which is missing from current combinatorial models.
Additionally, existing models lack a mechanism for handling noise in datasets.
We address these concerns by generalizing Edge-Colored Clustering, a recent
framework for categorical clustering of hypergraphs. Our generalizations allow
for a budgeted number of either (a) overlapping cluster assignments or (b) node
deletions. For each new model we present a greedy algorithm which approximately
minimizes an edge mistake objective, as well as bicriteria approximations where
the second approximation factor is on the budget. Additionally, we address the
parameterized complexity of each problem, providing FPT algorithms and hardness
results
Towards optimal symbolization for time series comparisons
The abundance and value of mining large time series data sets has long been acknowledged. Ubiquitous in fields ranging from astronomy, biology and web science the size and number of these datasets continues to increase, a situation exacerbated by the exponential growth of our digital footprints. The prevalence and potential utility of this data has led to a vast number of time-series data mining techniques, many of which require symbolization of the raw time series as a pre-processing step for which a number of well used, pre-existing approaches from the literature are typically employed. In this work we note that these standard approaches are sub-optimal in (at least) the broad application area of time series comparison leading to unnecessary data corruption and potential performance loss before any real data mining takes place. Addressing this we present a novel quantizer based upon optimization of comparison fidelity and a computationally tractable algorithm for its implementation on big datasets. We demonstrate empirically that our new approach provides a statistically significant reduction in the amount of error introduced by the symbolization process compared to current state-of-the-art. The approach therefore provides a more accurate input for the vast number of data mining techniques in the literature, providing the potential of increased real world performance across a wide range of existing data mining algorithms and applications
- …