2,974 research outputs found
Graviton-Photon Conversion in Atoms and the Detection of Gravitons
We study graviton-photon conversion in ground-based experiments. From
graviton to photon tran- sition, we calculate the cross section of
graviton-atom interaction in the presence of spherical atomic electric fields;
the obtained results hold for graviton energy around 100 keV - 1 GeV and would
be enhanced by crystal structures, thus it gives a chance to catch MeV level
gravitons from the universe with current neutrino facilities. From photon to
graviton transition, we propose an experiment using entangled pho- ton pairs to
count missing photons passing through transverse magnetic tunnel, which could
be used to verify the energy quantization of gravitational field
A note on the fractional Cauchy problems with nonlocal initial conditions
AbstractOf concern is the Cauchy problems for fractional integro-differential equations with nonlocal initial conditions. Using a new strategy in terms of the compactness of the semigroup generated by the operator in the linear part and approximating technique, a new existence theorem for mild solutions is established. An application to a fractional partial integro-differential equation with a nonlocal initial condition is also considered
One Fits All:Power General Time Series Analysis by Pretrained LM
Although we have witnessed great success of pre-trained models in natural
language processing (NLP) and computer vision (CV), limited progress has been
made for general time series analysis. Unlike NLP and CV where a unified model
can be used to perform different tasks, specially designed approach still
dominates in each time series analysis task such as classification, anomaly
detection, forecasting, and few-shot learning. The main challenge that blocks
the development of pre-trained model for time series analysis is the lack of a
large amount of data for training. In this work, we address this challenge by
leveraging language or CV models, pre-trained from billions of tokens, for time
series analysis. Specifically, we refrain from altering the self-attention and
feedforward layers of the residual blocks in the pre-trained language or image
model. This model, known as the Frozen Pretrained Transformer (FPT), is
evaluated through fine-tuning on all major types of tasks involving time
series. Our results demonstrate that pre-trained models on natural language or
images can lead to a comparable or state-of-the-art performance in all main
time series analysis tasks, as illustrated in Figure 1. We also found both
theoretically and empirically that the self-attention module behaviors
similarly to principle component analysis (PCA), an observation that helps
explains how transformer bridges the domain gap and a crucial step towards
understanding the universality of a pre-trained transformer.The code is
publicly available at https://github.com/DAMO-DI-ML/One_Fits_All.Comment: Neurips 2023 Spotligh
One Fits All: Universal Time Series Analysis by Pretrained LM and Specially Designed Adaptors
Despite the impressive achievements of pre-trained models in the fields of
natural language processing (NLP) and computer vision (CV), progress in the
domain of time series analysis has been limited. In contrast to NLP and CV,
where a single model can handle various tasks, time series analysis still
relies heavily on task-specific methods for activities such as classification,
anomaly detection, forecasting, and few-shot learning. The primary obstacle to
developing a pre-trained model for time series analysis is the scarcity of
sufficient training data. In our research, we overcome this obstacle by
utilizing pre-trained models from language or CV, which have been trained on
billions of data points, and apply them to time series analysis. We assess the
effectiveness of the pre-trained transformer model in two ways. Initially, we
maintain the original structure of the self-attention and feedforward layers in
the residual blocks of the pre-trained language or image model, using the
Frozen Pre-trained Transformer (FPT) for time series analysis with the addition
of projection matrices for input and output. Additionally, we introduce four
unique adapters, designed specifically for downstream tasks based on the
pre-trained model, including forecasting and anomaly detection. These adapters
are further enhanced with efficient parameter tuning, resulting in superior
performance compared to all state-of-the-art methods.Our comprehensive
experimental studies reveal that (a) the simple FPT achieves top-tier
performance across various time series analysis tasks; and (b) fine-tuning the
FPT with the custom-designed adapters can further elevate its performance,
outshining specialized task-specific models.Comment: this article draws heavily from arXiv:2302.1193
SVQ: Sparse Vector Quantization for Spatiotemporal Forecasting
Spatio-temporal forecasting, pivotal in numerous fields, hinges on the
delicate equilibrium between isolating nuanced patterns and sifting out noise.
To tackle this, we introduce Sparse Regression-based Vector Quantization (SVQ),
a novel technique that leverages sparse regression for succinct representation,
an approach theoretically and practically favored over classical
clustering-based vector quantization methods. This approach preserves critical
details from the original vectors using a regression model while filtering out
noise via sparse design. Moreover, we approximate the sparse regression process
using a blend of a two-layer MLP and an extensive codebook. This approach not
only substantially cuts down on computational costs but also grants SVQ
differentiability and training simplicity, resulting in a notable enhancement
of performance. Our empirical studies on five spatial-temporal benchmark
datasets demonstrate that SVQ achieves state-of-the-art results. Specifically,
on the WeatherBench-S temperature dataset, SVQ improves the top baseline by
7.9%. In video prediction benchmarks-Human, KTH, and KittiCaltech-it reduces
MAE by an average of 9.4% and improves image quality by 17.3% (LPIPS)
OneNet: Enhancing Time Series Forecasting Models under Concept Drift by Online Ensembling
Online updating of time series forecasting models aims to address the concept
drifting problem by efficiently updating forecasting models based on streaming
data. Many algorithms are designed for online time series forecasting, with
some exploiting cross-variable dependency while others assume independence
among variables. Given every data assumption has its own pros and cons in
online time series modeling, we propose \textbf{On}line \textbf{e}nsembling
\textbf{Net}work (OneNet). It dynamically updates and combines two models, with
one focusing on modeling the dependency across the time dimension and the other
on cross-variate dependency. Our method incorporates a reinforcement
learning-based approach into the traditional online convex programming
framework, allowing for the linear combination of the two models with
dynamically adjusted weights. OneNet addresses the main shortcoming of
classical online learning methods that tend to be slow in adapting to the
concept drift. Empirical results show that OneNet reduces online forecasting
error by more than compared to the State-Of-The-Art (SOTA)
method. The code is available at \url{https://github.com/yfzhang114/OneNet}.Comment: 32 pages, 11 figures, 37th Conference on Neural Information
Processing Systems (NeurIPS 2023
- …