1,185 research outputs found
A Survey on Temporal Knowledge Graph Completion: Taxonomy, Progress, and Prospects
Temporal characteristics are prominently evident in a substantial volume of
knowledge, which underscores the pivotal role of Temporal Knowledge Graphs
(TKGs) in both academia and industry. However, TKGs often suffer from
incompleteness for three main reasons: the continuous emergence of new
knowledge, the weakness of the algorithm for extracting structured information
from unstructured data, and the lack of information in the source dataset.
Thus, the task of Temporal Knowledge Graph Completion (TKGC) has attracted
increasing attention, aiming to predict missing items based on the available
information. In this paper, we provide a comprehensive review of TKGC methods
and their details. Specifically, this paper mainly consists of three
components, namely, 1)Background, which covers the preliminaries of TKGC
methods, loss functions required for training, as well as the dataset and
evaluation protocol; 2)Interpolation, that estimates and predicts the missing
elements or set of elements through the relevant available information. It
further categorizes related TKGC methods based on how to process temporal
information; 3)Extrapolation, which typically focuses on continuous TKGs and
predicts future events, and then classifies all extrapolation methods based on
the algorithms they utilize. We further pinpoint the challenges and discuss
future research directions of TKGC
Challenges in Complex Systems Science
FuturICT foundations are social science, complex systems science, and ICT.
The main concerns and challenges in the science of complex systems in the
context of FuturICT are laid out in this paper with special emphasis on the
Complex Systems route to Social Sciences. This include complex systems having:
many heterogeneous interacting parts; multiple scales; complicated transition
laws; unexpected or unpredicted emergence; sensitive dependence on initial
conditions; path-dependent dynamics; networked hierarchical connectivities;
interaction of autonomous agents; self-organisation; non-equilibrium dynamics;
combinatorial explosion; adaptivity to changing environments; co-evolving
subsystems; ill-defined boundaries; and multilevel dynamics. In this context,
science is seen as the process of abstracting the dynamics of systems from
data. This presents many challenges including: data gathering by large-scale
experiment, participatory sensing and social computation, managing huge
distributed dynamic and heterogeneous databases; moving from data to dynamical
models, going beyond correlations to cause-effect relationships, understanding
the relationship between simple and comprehensive models with appropriate
choices of variables, ensemble modeling and data assimilation, modeling systems
of systems of systems with many levels between micro and macro; and formulating
new approaches to prediction, forecasting, and risk, especially in systems that
can reflect on and change their behaviour in response to predictions, and
systems whose apparently predictable behaviour is disrupted by apparently
unpredictable rare or extreme events. These challenges are part of the FuturICT
agenda
From Data to Actions in Intelligent Transportation Systems: A Prescription of Functional Requirements for Model Actionability
Advances in Data Science permeate every field of Transportation Science and Engineering,
resulting in developments in the transportation sector that are data-driven. Nowadays, Intelligent
Transportation Systems (ITS) could be arguably approached as a “story” intensively producing and
consuming large amounts of data. A diversity of sensing devices densely spread over the infrastructure,
vehicles or the travelers’ personal devices act as sources of data flows that are eventually
fed into software running on automatic devices, actuators or control systems producing, in turn,
complex information flows among users, traffic managers, data analysts, traffic modeling scientists,
etc. These information flows provide enormous opportunities to improve model development and
decision-making. This work aims to describe how data, coming from diverse ITS sources, can be used
to learn and adapt data-driven models for efficiently operating ITS assets, systems and processes;
in other words, for data-based models to fully become actionable. Grounded in this described data
modeling pipeline for ITS, we define the characteristics, engineering requisites and challenges intrinsic
to its three compounding stages, namely, data fusion, adaptive learning and model evaluation.
We deliberately generalize model learning to be adaptive, since, in the core of our paper is the firm
conviction that most learners will have to adapt to the ever-changing phenomenon scenario underlying
the majority of ITS applications. Finally, we provide a prospect of current research lines within
Data Science that can bring notable advances to data-based ITS modeling, which will eventually
bridge the gap towards the practicality and actionability of such models.This work was supported in part by the Basque Government for its funding support through the EMAITEK program (3KIA, ref. KK-2020/00049). It has also received funding support from the Consolidated Research Group MATHMODE (IT1294-19) granted by the Department of Education of the Basque Government
Data Augmentation techniques in time series domain: A survey and taxonomy
With the latest advances in Deep Learning-based generative models, it has not
taken long to take advantage of their remarkable performance in the area of
time series. Deep neural networks used to work with time series heavily depend
on the size and consistency of the datasets used in training. These features
are not usually abundant in the real world, where they are usually limited and
often have constraints that must be guaranteed. Therefore, an effective way to
increase the amount of data is by using Data Augmentation techniques, either by
adding noise or permutations and by generating new synthetic data. This work
systematically reviews the current state-of-the-art in the area to provide an
overview of all available algorithms and proposes a taxonomy of the most
relevant research. The efficiency of the different variants will be evaluated
as a central part of the process, as well as the different metrics to evaluate
the performance and the main problems concerning each model will be analysed.
The ultimate aim of this study is to provide a summary of the evolution and
performance of areas that produce better results to guide future researchers in
this field.Comment: 33 pages, 9 figure
2023 SDSU Data Science Symposium Presentation Abstracts
This document contains abstracts for presentations and posters 2023 SDSU Data Science Symposium
- …