35 research outputs found
Self-similarity parameter estimation and reproduction property for non-Gaussian Hermite processes
We consider the class of all the Hermite processes of order and with Hurst parameter . The process is -selfsimilar, it has
stationary increments and it exhibits long-range dependence identical to that
of fractional Brownian motion (fBm). For , is fBm, which is
Gaussian; for , is the Rosenblatt process, which lives in the
second Wiener chaos; for any , is a process in the th
Wiener chaos. We study the variations of for any , by using
multiple Wiener -It\^{o} stochastic integrals and Malliavin calculus. We prove
a reproduction property for this class of processes in the sense that the terms
appearing in the chaotic decomposition of their variations give rise to other
Hermite processes of different orders and with different Hurst parameters. We
apply our results to construct a strongly consistent estimator for the
self-similarity parameter from discrete observations of ; the
asymptotics of this estimator, after appropriate normalization, are proved to
be distributed like a Rosenblatt random variable (value at time of a
Rosenblatt process).with self-similarity parameter .Comment: To appear in "Communications on Stochastic Analysis
Mitigating Data Imbalance and Representation Degeneration in Multilingual Machine Translation
Despite advances in multilingual neural machine translation (MNMT), we argue
that there are still two major challenges in this area: data imbalance and
representation degeneration. The data imbalance problem refers to the imbalance
in the amount of parallel corpora for all language pairs, especially for
long-tail languages (i.e., very low-resource languages). The representation
degeneration problem refers to the problem of encoded tokens tending to appear
only in a small subspace of the full space available to the MNMT model. To
solve these two issues, we propose Bi-ACL, a framework that uses only
target-side monolingual data and a bilingual dictionary to improve the
performance of the MNMT model. We define two modules, named bidirectional
autoencoder and bidirectional contrastive learning, which we combine with an
online constrained beam search and a curriculum learning sampling strategy.
Extensive experiments show that our proposed method is more effective both in
long-tail languages and in high-resource languages. We also demonstrate that
our approach is capable of transferring knowledge between domains and languages
in zero-shot scenarios.Comment: Accepted to Findings of EMNLP 2023, add statistical significance
tests. code available at https://github.com/lavine-lmu/Bi-AC