10 research outputs found
Domain Adaptation for Time Series Forecasting via Attention Sharing
Recent years have witnessed deep neural networks gaining increasing
popularity in the field of time series forecasting. A primary reason of their
success is their ability to effectively capture complex temporal dynamics
across multiple related time series. However, the advantages of these deep
forecasters only start to emerge in the presence of a sufficient amount of
data. This poses a challenge for typical forecasting problems in practice,
where one either has a small number of time series, or limited observations per
time series, or both. To cope with the issue of data scarcity, we propose a
novel domain adaptation framework, Domain Adaptation Forecaster (DAF), that
leverages the statistical strengths from another relevant domain with abundant
data samples (source) to improve the performance on the domain of interest with
limited data (target). In particular, we propose an attention-based shared
module with a domain discriminator across domains as well as private modules
for individual domains. This allows us to jointly train the source and target
domains by generating domain-invariant latent features while retraining
domain-specific features. Extensive experiments on various domains demonstrate
that our proposed method outperforms state-of-the-art baselines on synthetic
and real-world datasets.Comment: 19 pages, 9 figure
Theoretical Guarantees of Learning Ensembling Strategies with Applications to Time Series Forecasting
Ensembling is among the most popular tools in machine learning (ML) due to
its effectiveness in minimizing variance and thus improving generalization.
Most ensembling methods for black-box base learners fall under the umbrella of
"stacked generalization," namely training an ML algorithm that takes the
inferences from the base learners as input. While stacking has been widely
applied in practice, its theoretical properties are poorly understood. In this
paper, we prove a novel result, showing that choosing the best stacked
generalization from a (finite or finite-dimensional) family of stacked
generalizations based on cross-validated performance does not perform "much
worse" than the oracle best. Our result strengthens and significantly extends
the results in Van der Laan et al. (2007). Inspired by the theoretical
analysis, we further propose a particular family of stacked generalizations in
the context of probabilistic forecasting, each one with a different sensitivity
for how much the ensemble weights are allowed to vary across items, timestamps
in the forecast horizon, and quantiles. Experimental results demonstrate the
performance gain of the proposed method.Comment: ICML 202
Learning Physical Models that Can Respect Conservation Laws
Recent work in scientific machine learning (SciML) has focused on
incorporating partial differential equation (PDE) information into the learning
process. Much of this work has focused on relatively ``easy'' PDE operators
(e.g., elliptic and parabolic), with less emphasis on relatively ``hard'' PDE
operators (e.g., hyperbolic). Within numerical PDEs, the latter problem class
requires control of a type of volume element or conservation constraint, which
is known to be challenging. Delivering on the promise of SciML requires
seamlessly incorporating both types of problems into the learning process. To
address this issue, we propose ProbConserv, a framework for incorporating
conservation constraints into a generic SciML architecture. To do so,
ProbConserv combines the integral form of a conservation law with a Bayesian
update. We provide a detailed analysis of ProbConserv on learning with the
Generalized Porous Medium Equation (GPME), a widely-applicable parameterized
family of PDEs that illustrates the qualitative properties of both easier and
harder PDEs. ProbConserv is effective for easy GPME variants, performing well
with state-of-the-art competitors; and for harder GPME variants it outperforms
other approaches that do not guarantee volume conservation. ProbConserv
seamlessly enforces physical conservation constraints, maintains probabilistic
uncertainty quantification (UQ), and deals well with shocks and
heteroscedasticities. In each case, it achieves superior predictive performance
on downstream tasks
PreDiff: Precipitation Nowcasting with Latent Diffusion Models
Earth system forecasting has traditionally relied on complex physical models
that are computationally expensive and require significant domain expertise. In
the past decade, the unprecedented increase in spatiotemporal Earth observation
data has enabled data-driven forecasting models using deep learning techniques.
These models have shown promise for diverse Earth system forecasting tasks but
either struggle with handling uncertainty or neglect domain-specific prior
knowledge, resulting in averaging possible futures to blurred forecasts or
generating physically implausible predictions. To address these limitations, we
propose a two-stage pipeline for probabilistic spatiotemporal forecasting: 1)
We develop PreDiff, a conditional latent diffusion model capable of
probabilistic forecasts. 2) We incorporate an explicit knowledge control
mechanism to align forecasts with domain-specific physical constraints. This is
achieved by estimating the deviation from imposed constraints at each denoising
step and adjusting the transition distribution accordingly. We conduct
empirical studies on two datasets: N-body MNIST, a synthetic dataset with
chaotic behavior, and SEVIR, a real-world precipitation nowcasting dataset.
Specifically, we impose the law of conservation of energy in N-body MNIST and
anticipated precipitation intensity in SEVIR. Experiments demonstrate the
effectiveness of PreDiff in handling uncertainty, incorporating domain-specific
prior knowledge, and generating forecasts that exhibit high operational
utility.Comment: Technical repor
Neural forecasting: Introduction and literature overview
Neural network based forecasting methods have become ubiquitous in
large-scale industrial forecasting applications over the last years. As the
prevalence of neural network based solutions among the best entries in the
recent M4 competition shows, the recent popularity of neural forecasting
methods is not limited to industry and has also reached academia. This article
aims at providing an introduction and an overview of some of the advances that
have permitted the resurgence of neural networks in machine learning. Building
on these foundations, the article then gives an overview of the recent
literature on neural networks for forecasting and applications.Comment: 66 pages, 5 figure