661 research outputs found
Object Detection in 20 Years: A Survey
Object detection, as of one the most fundamental and challenging problems in
computer vision, has received great attention in recent years. Its development
in the past two decades can be regarded as an epitome of computer vision
history. If we think of today's object detection as a technical aesthetics
under the power of deep learning, then turning back the clock 20 years we would
witness the wisdom of cold weapon era. This paper extensively reviews 400+
papers of object detection in the light of its technical evolution, spanning
over a quarter-century's time (from the 1990s to 2019). A number of topics have
been covered in this paper, including the milestone detectors in history,
detection datasets, metrics, fundamental building blocks of the detection
system, speed up techniques, and the recent state of the art detection methods.
This paper also reviews some important detection applications, such as
pedestrian detection, face detection, text detection, etc, and makes an in-deep
analysis of their challenges as well as technical improvements in recent years.Comment: This work has been submitted to the IEEE TPAMI for possible
publicatio
Least Squares Based and Two-Stage Least Squares Based Iterative Estimation Algorithms for H-FIR-MA Systems
This paper studies the identification of Hammerstein finite impulse response moving average (H-FIR-MA for short) systems. A new two-stage least squares iterative algorithm is developed to identify the parameters of the H-FIR-MA systems. The simulation cases indicate the efficiency of the proposed algorithms
Learning from big imaging data to predict radiotherapy treatment outcomes and side-effects
The prevalence of cancer is an increasing healthcare issue as it is the predominant cause of death worldwide. The growing cancer burden is caused by several factors including population growth, aging, and the changing prevalence of certain causes of cancer during social and economic development. To address the global cancer burden, new technologies, for instance Artificial Intelligence (AI), have been applied in the workflow of cancer care from diagnosis to treatment. For cancer treatment, especially radiotherapy, new innovations are not only useful to provide comprehensive treatment plans, but also able to reduce radiotherapy- induced side-effects which may exist in patients during and (long) after treatment. This thesis focuses on AI-based quantitative imaging techniques (e.g., radiomics) that have the potential to assist doctors and patients to make individualized treatment decisions
Bayesian Domain Invariant Learning via Posterior Generalization of Parameter Distributions
Domain invariant learning aims to learn models that extract invariant
features over various training domains, resulting in better generalization to
unseen target domains. Recently, Bayesian Neural Networks have achieved
promising results in domain invariant learning, but most works concentrate on
aligning features distributions rather than parameter distributions. Inspired
by the principle of Bayesian Neural Network, we attempt to directly learn the
domain invariant posterior distribution of network parameters. We first propose
a theorem to show that the invariant posterior of parameters can be implicitly
inferred by aggregating posteriors on different training domains. Our
assumption is more relaxed and allows us to extract more domain invariant
information. We also propose a simple yet effective method, named PosTerior
Generalization (PTG), that can be used to estimate the invariant parameter
distribution. PTG fully exploits variational inference to approximate parameter
distributions, including the invariant posterior and the posteriors on training
domains. Furthermore, we develop a lite version of PTG for widespread
applications. PTG shows competitive performance on various domain
generalization benchmarks on DomainBed. Additionally, PTG can use any existing
domain generalization methods as its prior, and combined with previous
state-of-the-art method the performance can be further improved. Code will be
made public
Dual-Branched Spatio-temporal Fusion Network for Multi-horizon Tropical Cyclone Track Forecast
Tropical cyclone (TC) is an extreme tropical weather system and its
trajectory can be described by a variety of spatio-temporal data. Effective
mining of these data is the key to accurate TCs track forecasting. However,
existing methods face the problem that the model complexity is too high or it
is difficult to efficiently extract features from multi-modal data. In this
paper, we propose the Dual-Branched spatio-temporal Fusion Network (DBF-Net) --
a novel multi-horizon tropical cyclone track forecasting model which fuses the
multi-modal features efficiently. DBF-Net contains a TC features branch that
extracts temporal features from 1D inherent features of TCs and a pressure
field branch that extracts spatio-temporal features from reanalysis 2D pressure
field. Through the encoder-decoder-based architecture and efficient feature
fusion, DBF-Net can fully mine the information of the two types of data, and
achieve good TCs track prediction results. Extensive experiments on historical
TCs track data in the Northwest Pacific show that our DBF-Net achieves
significant improvement compared with existing statistical and deep learning
TCs track forecast methods
Semi-blind source extraction algorithm for fetal electrocardiogram based on generalized autocorrelations and reference signals
AbstractBlind source extraction (BSE) has become one of the promising methods in the field of signal processing and analysis, which only desires to extract “interesting” source signals with specific stochastic property or features so as to save lots of computing time and resources. This paper addresses BSE problem, in which desired source signals have some available reference signals. Based on this prior information, we develop an objective function for extraction of temporally correlated sources. Maximizing this objective function, a semi-blind source extraction fixed-point algorithm is proposed. Simulations on artificial electrocardiograph (ECG) signals and the real-world ECG data demonstrate the better performance of the new algorithm. Moreover, comparisons with existing algorithms further indicate the validity of our new algorithm, and also show its robustness to the estimated error of time delay
SRDA-Net: Super-Resolution Domain Adaptation Networks for Semantic Segmentation
Recently, Unsupervised Domain Adaptation was proposed to address the domain
shift problem in semantic segmentation task, but it may perform poor when
source and target domains belong to different resolutions. In this work, we
design a novel end-to-end semantic segmentation network, Super-Resolution
Domain Adaptation Network (SRDA-Net), which could simultaneously complete
super-resolution and domain adaptation. Such characteristics exactly meet the
requirement of semantic segmentation for remote sensing images which usually
involve various resolutions. Generally, SRDA-Net includes three deep neural
networks: a Super-Resolution and Segmentation (SRS) model focuses on recovering
high-resolution image and predicting segmentation map; a pixel-level domain
classifier (PDC) tries to distinguish the images from which domains; and
output-space domain classifier (ODC) discriminates pixel label distributions
from which domains. PDC and ODC are considered as the discriminators, and SRS
is treated as the generator. By the adversarial learning, SRS tries to align
the source with target domains on pixel-level visual appearance and
output-space. Experiments are conducted on the two remote sensing datasets with
different resolutions. SRDA-Net performs favorably against the state-of-the-art
methods in terms of accuracy and visual quality. Code and models are available
at https://github.com/tangzhenjie/SRDA-Net
- …