849 research outputs found
Disentangled Representation Learning
Disentangled Representation Learning (DRL) aims to learn a model capable of
identifying and disentangling the underlying factors hidden in the observable
data in representation form. The process of separating underlying factors of
variation into variables with semantic meaning benefits in learning explainable
representations of data, which imitates the meaningful understanding process of
humans when observing an object or relation. As a general learning strategy,
DRL has demonstrated its power in improving the model explainability,
controlability, robustness, as well as generalization capacity in a wide range
of scenarios such as computer vision, natural language processing, data mining
etc. In this article, we comprehensively review DRL from various aspects
including motivations, definitions, methodologies, evaluations, applications
and model designs. We discuss works on DRL based on two well-recognized
definitions, i.e., Intuitive Definition and Group Theory Definition. We further
categorize the methodologies for DRL into four groups, i.e., Traditional
Statistical Approaches, Variational Auto-encoder Based Approaches, Generative
Adversarial Networks Based Approaches, Hierarchical Approaches and Other
Approaches. We also analyze principles to design different DRL models that may
benefit different tasks in practical applications. Finally, we point out
challenges in DRL as well as potential research directions deserving future
investigations. We believe this work may provide insights for promoting the DRL
research in the community.Comment: 22 pages,9 figure
To Compress or Not to Compress -- Self-Supervised Learning and Information Theory: A Review
Deep neural networks have demonstrated remarkable performance in supervised
learning tasks but require large amounts of labeled data. Self-supervised
learning offers an alternative paradigm, enabling the model to learn from data
without explicit labels. Information theory has been instrumental in
understanding and optimizing deep neural networks. Specifically, the
information bottleneck principle has been applied to optimize the trade-off
between compression and relevant information preservation in supervised
settings. However, the optimal information objective in self-supervised
learning remains unclear. In this paper, we review various approaches to
self-supervised learning from an information-theoretic standpoint and present a
unified framework that formalizes the \textit{self-supervised
information-theoretic learning problem}. We integrate existing research into a
coherent framework, examine recent self-supervised methods, and identify
research opportunities and challenges. Moreover, we discuss empirical
measurement of information-theoretic quantities and their estimators. This
paper offers a comprehensive review of the intersection between information
theory, self-supervised learning, and deep neural networks
A stochastic-variational model for soft Mumford-Shah segmentation
In contemporary image and vision analysis, stochastic approaches demonstrate
great flexibility in representing and modeling complex phenomena, while
variational-PDE methods gain enormous computational advantages over Monte-Carlo
or other stochastic algorithms. In combination, the two can lead to much more
powerful novel models and efficient algorithms. In the current work, we propose
a stochastic-variational model for soft (or fuzzy) Mumford-Shah segmentation of
mixture image patterns. Unlike the classical hard Mumford-Shah segmentation,
the new model allows each pixel to belong to each image pattern with some
probability. We show that soft segmentation leads to hard segmentation, and
hence is more general. The modeling procedure, mathematical analysis, and
computational implementation of the new model are explored in detail, and
numerical examples of synthetic and natural images are presented.Comment: 22 page
Transfer Learning for Speech and Language Processing
Transfer learning is a vital technique that generalizes models trained for
one setting or task to other settings or tasks. For example in speech
recognition, an acoustic model trained for one language can be used to
recognize speech in another language, with little or no re-training data.
Transfer learning is closely related to multi-task learning (cross-lingual vs.
multilingual), and is traditionally studied in the name of `model adaptation'.
Recent advance in deep learning shows that transfer learning becomes much
easier and more effective with high-level abstract features learned by deep
models, and the `transfer' can be conducted not only between data distributions
and data types, but also between model structures (e.g., shallow nets and deep
nets) or even model types (e.g., Bayesian models and neural models). This
review paper summarizes some recent prominent research towards this direction,
particularly for speech and language processing. We also report some results
from our group and highlight the potential of this very interesting research
field.Comment: 13 pages, APSIPA 201
IoT Anomaly Detection Methods and Applications: A Survey
Ongoing research on anomaly detection for the Internet of Things (IoT) is a
rapidly expanding field. This growth necessitates an examination of application
trends and current gaps. The vast majority of those publications are in areas
such as network and infrastructure security, sensor monitoring, smart home, and
smart city applications and are extending into even more sectors. Recent
advancements in the field have increased the necessity to study the many IoT
anomaly detection applications. This paper begins with a summary of the
detection methods and applications, accompanied by a discussion of the
categorization of IoT anomaly detection algorithms. We then discuss the current
publications to identify distinct application domains, examining papers chosen
based on our search criteria. The survey considers 64 papers among recent
publications published between January 2019 and July 2021. In recent
publications, we observed a shortage of IoT anomaly detection methodologies,
for example, when dealing with the integration of systems with various sensors,
data and concept drifts, and data augmentation where there is a shortage of
Ground Truth data. Finally, we discuss the present such challenges and offer
new perspectives where further research is required.Comment: 22 page
- …