46 research outputs found
Time-Series Contrastive Learning against False Negatives and Class Imbalance
As an exemplary self-supervised approach for representation learning,
time-series contrastive learning has exhibited remarkable advancements in
contemporary research. While recent contrastive learning strategies have
focused on how to construct appropriate positives and negatives, in this study,
we conduct theoretical analysis and find they have overlooked the fundamental
issues: false negatives and class imbalance inherent in the InfoNCE loss-based
framework. Therefore, we introduce a straightforward modification grounded in
the SimCLR framework, universally adaptable to models engaged in the instance
discrimination task. By constructing instance graphs to facilitate interactive
learning among instances, we emulate supervised contrastive learning via the
multiple-instances discrimination task, mitigating the harmful impact of false
negatives. Moreover, leveraging the graph structure and few-labeled data, we
perform semi-supervised consistency classification and enhance the
representative ability of minority classes. We compared our method with the
most popular time-series contrastive learning methods on four real-world
time-series datasets and demonstrated our significant advantages in overall
performance
Origin-Destination Travel Time Oracle for Map-based Services
Given an origin (O), a destination (D), and a departure time (T), an
Origin-Destination (OD) travel time oracle~(ODT-Oracle) returns an estimate of
the time it takes to travel from O to D when departing at T. ODT-Oracles serve
important purposes in map-based services. To enable the construction of such
oracles, we provide a travel-time estimation (TTE) solution that leverages
historical trajectories to estimate time-varying travel times for OD pairs.
The problem is complicated by the fact that multiple historical trajectories
with different travel times may connect an OD pair, while trajectories may vary
from one another. To solve the problem, it is crucial to remove outlier
trajectories when doing travel time estimation for future queries.
We propose a novel, two-stage framework called Diffusion-based
Origin-destination Travel Time Estimation (DOT), that solves the problem.
First, DOT employs a conditioned Pixelated Trajectories (PiT) denoiser that
enables building a diffusion-based PiT inference process by learning
correlations between OD pairs and historical trajectories. Specifically, given
an OD pair and a departure time, we aim to infer a PiT. Next, DOT encompasses a
Masked Vision Transformer~(MViT) that effectively and efficiently estimates a
travel time based on the inferred PiT. We report on extensive experiments on
two real-world datasets that offer evidence that DOT is capable of
outperforming baseline methods in terms of accuracy, scalability, and
explainability.Comment: 15 pages, 12 figures, accepted by SIGMOD International Conference on
Management of Data 202
Look Beneath the Surface: Exploiting Fundamental Symmetry for Sample-Efficient Offline RL
Offline reinforcement learning (RL) offers an appealing approach to
real-world tasks by learning policies from pre-collected datasets without
interacting with the environment. However, the performance of existing offline
RL algorithms heavily depends on the scale and state-action space coverage of
datasets. Real-world data collection is often expensive and uncontrollable,
leading to small and narrowly covered datasets and posing significant
challenges for practical deployments of offline RL. In this paper, we provide a
new insight that leveraging the fundamental symmetry of system dynamics can
substantially enhance offline RL performance under small datasets.
Specifically, we propose a Time-reversal symmetry (T-symmetry) enforced
Dynamics Model (TDM), which establishes consistency between a pair of forward
and reverse latent dynamics. TDM provides both well-behaved representations for
small datasets and a new reliability measure for OOD samples based on
compliance with the T-symmetry. These can be readily used to construct a new
offline RL algorithm (TSRL) with less conservative policy constraints and a
reliable latent space data augmentation procedure. Based on extensive
experiments, we find TSRL achieves great performance on small benchmark
datasets with as few as 1% of the original samples, which significantly
outperforms the recent offline RL algorithms in terms of data efficiency and
generalizability.Comment: The first two authors contributed equall
A Survey on Service Route and Time Prediction in Instant Delivery: Taxonomy, Progress, and Prospects
Instant delivery services, such as food delivery and package delivery, have
achieved explosive growth in recent years by providing customers with
daily-life convenience. An emerging research area within these services is
service Route\&Time Prediction (RTP), which aims to estimate the future service
route as well as the arrival time of a given worker. As one of the most crucial
tasks in those service platforms, RTP stands central to enhancing user
satisfaction and trimming operational expenditures on these platforms. Despite
a plethora of algorithms developed to date, there is no systematic,
comprehensive survey to guide researchers in this domain. To fill this gap, our
work presents the first comprehensive survey that methodically categorizes
recent advances in service route and time prediction. We start by defining the
RTP challenge and then delve into the metrics that are often employed.
Following that, we scrutinize the existing RTP methodologies, presenting a
novel taxonomy of them. We categorize these methods based on three criteria:
(i) type of task, subdivided into only-route prediction, only-time prediction,
and joint route\&time prediction; (ii) model architecture, which encompasses
sequence-based and graph-based models; and (iii) learning paradigm, including
Supervised Learning (SL) and Deep Reinforcement Learning (DRL). Conclusively,
we highlight the limitations of current research and suggest prospective
avenues. We believe that the taxonomy, progress, and prospects introduced in
this paper can significantly promote the development of this field
Attention-based Multi-Level Fusion Network for Light Field Depth Estimation
Depth estimation from Light Field (LF) images is a crucial basis for LF related applications. Since multiple views with abundant information are available, how to effectively fuse features of these views is a key point for accurate LF depth estimation. In this paper, we propose a novel attention-based multi-level fusion network. Combined with the four-branch structure, we design intra-branch fusion strategy and inter-branch fusion strategy to hierarchically fuse effective features from different views. By introducing the attention mechanism, features of views with less occlusions and richer textures are selected inside and between these branches to provide more effective information for depth estimation. The depth maps are finally estimated after further aggregation. Experimental results shows the proposed method achieves state-of-the-art performance in both quantitative and qualitative evaluation, which also ranks first in the commonly used HCI 4D Light Field Benchmark