410,467 research outputs found
Certifiably Robust Interpretation in Deep Learning
Deep learning interpretation is essential to explain the reasoning behind
model predictions. Understanding the robustness of interpretation methods is
important especially in sensitive domains such as medical applications since
interpretation results are often used in downstream tasks. Although
gradient-based saliency maps are popular methods for deep learning
interpretation, recent works show that they can be vulnerable to adversarial
attacks. In this paper, we address this problem and provide a certifiable
defense method for deep learning interpretation. We show that a sparsified
version of the popular SmoothGrad method, which computes the average saliency
maps over random perturbations of the input, is certifiably robust against
adversarial perturbations. We obtain this result by extending recent bounds for
certifiably robust smooth classifiers to the interpretation setting.
Experiments on ImageNet samples validate our theory
Model-Based Robust Deep Learning: Generalizing to Natural, Out-of-Distribution Data
While deep learning has resulted in major breakthroughs in many application
domains, the frameworks commonly used in deep learning remain fragile to
artificially-crafted and imperceptible changes in the data. In response to this
fragility, adversarial training has emerged as a principled approach for
enhancing the robustness of deep learning with respect to norm-bounded
perturbations. However, there are other sources of fragility for deep learning
that are arguably more common and less thoroughly studied. Indeed, natural
variation such as lighting or weather conditions can significantly degrade the
accuracy of trained neural networks, proving that such natural variation
presents a significant challenge for deep learning.
In this paper, we propose a paradigm shift from perturbation-based
adversarial robustness toward model-based robust deep learning. Our objective
is to provide general training algorithms that can be used to train deep neural
networks to be robust against natural variation in data. Critical to our
paradigm is first obtaining a model of natural variation which can be used to
vary data over a range of natural conditions. Such models may be either known a
priori or else learned from data. In the latter case, we show that deep
generative models can be used to learn models of natural variation that are
consistent with realistic conditions. We then exploit such models in three
novel model-based robust training algorithms in order to enhance the robustness
of deep learning with respect to the given model. Our extensive experiments
show that across a variety of naturally-occurring conditions and across various
datasets, deep neural networks trained with our model-based algorithms
significantly outperform both standard deep learning algorithms as well as
norm-bounded robust deep learning algorithms
Deep Robust Kalman Filter
A Robust Markov Decision Process (RMDP) is a sequential decision making model
that accounts for uncertainty in the parameters of dynamic systems. This
uncertainty introduces difficulties in learning an optimal policy, especially
for environments with large state spaces. We propose two algorithms, RTD-DQN
and Deep-RoK, for solving large-scale RMDPs using nonlinear approximation
schemes such as deep neural networks. The RTD-DQN algorithm incorporates the
robust Bellman temporal difference error into a robust loss function, yielding
robust policies for the agent. The Deep-RoK algorithm is a robust Bayesian
method, based on the Extended Kalman Filter (EKF), that accounts for both the
uncertainty in the weights of the approximated value function and the
uncertainty in the transition probabilities, improving the robustness of the
agent. We provide theoretical results for our approach and test the proposed
algorithms on a continuous state domain
Power of Deep Learning for Channel Estimation and Signal Detection in OFDM Systems
This article presents our initial results in deep learning for channel
estimation and signal detection in orthogonal frequency-division multiplexing
(OFDM). OFDM has been widely adopted in wireless broadband communications to
combat frequency-selective fading in wireless channels. In this article, we
take advantage of deep learning in handling wireless OFDM channels in an
end-to-end approach. Different from existing OFDM receivers that first estimate
CSI explicitly and then detect/recover the transmitted symbols with the
estimated CSI, our deep learning based approach estimates CSI implicitly and
recovers the transmitted symbols directly. To address channel distortion, a
deep learning model is first trained offline using the data generated from the
simulation based on the channel statistics and then used for recovering the
online transmitted data directly. From our simulation results, the deep
learning based approach has the ability to address channel distortions and
detect the transmitted symbols with performance comparable to minimum
mean-square error (MMSE) estimator. Furthermore, the deep learning based
approach is more robust than conventional methods when fewer training pilots
are used, the cyclic prefix (CP) is omitted, and nonlinear clipping noise is
presented. In summary, deep learning is a promising tool for channel estimation
and signal detection in wireless communications with complicated channel
distortions and interferences
Deep Forest
Current deep learning models are mostly build upon neural networks, i.e.,
multiple layers of parameterized differentiable nonlinear modules that can be
trained by backpropagation. In this paper, we explore the possibility of
building deep models based on non-differentiable modules. We conjecture that
the mystery behind the success of deep neural networks owes much to three
characteristics, i.e., layer-by-layer processing, in-model feature
transformation and sufficient model complexity. We propose the gcForest
approach, which generates \textit{deep forest} holding these characteristics.
This is a decision tree ensemble approach, with much less hyper-parameters than
deep neural networks, and its model complexity can be automatically determined
in a data-dependent way. Experiments show that its performance is quite robust
to hyper-parameter settings, such that in most cases, even across different
data from different domains, it is able to get excellent performance by using
the same default setting. This study opens the door of deep learning based on
non-differentiable modules, and exhibits the possibility of constructing deep
models without using backpropagation
End-to-End Optimization of Task-Oriented Dialogue Model with Deep Reinforcement Learning
In this paper, we present a neural network based task-oriented dialogue
system that can be optimized end-to-end with deep reinforcement learning (RL).
The system is able to track dialogue state, interface with knowledge bases, and
incorporate query results into agent's responses to successfully complete
task-oriented dialogues. Dialogue policy learning is conducted with a hybrid
supervised and deep RL methods. We first train the dialogue agent in a
supervised manner by learning directly from task-oriented dialogue corpora, and
further optimize it with deep RL during its interaction with users. In the
experiments on two different dialogue task domains, our model demonstrates
robust performance in tracking dialogue state and producing reasonable system
responses. We show that deep RL based optimization leads to significant
improvement on task success rate and reduction in dialogue length comparing to
supervised training model. We further show benefits of training task-oriented
dialogue model end-to-end comparing to component-wise optimization with
experiment results on dialogue simulations and human evaluations
Adversarial Attacks and Defences: A Survey
Deep learning has emerged as a strong and efficient framework that can be
applied to a broad spectrum of complex learning problems which were difficult
to solve using the traditional machine learning techniques in the past. In the
last few years, deep learning has advanced radically in such a way that it can
surpass human-level performance on a number of tasks. As a consequence, deep
learning is being extensively used in most of the recent day-to-day
applications. However, security of deep learning systems are vulnerable to
crafted adversarial examples, which may be imperceptible to the human eye, but
can lead the model to misclassify the output. In recent times, different types
of adversaries based on their threat model leverage these vulnerabilities to
compromise a deep learning system where adversaries have high incentives.
Hence, it is extremely important to provide robustness to deep learning
algorithms against these adversaries. However, there are only a few strong
countermeasures which can be used in all types of attack scenarios to design a
robust deep learning system. In this paper, we attempt to provide a detailed
discussion on different types of adversarial attacks with various threat models
and also elaborate the efficiency and challenges of recent countermeasures
against them
Security Consideration For Deep Learning-Based Image Forensics
Recently, image forensics community has paied attention to the research on
the design of effective algorithms based on deep learning technology and facts
proved that combining the domain knowledge of image forensics and deep learning
would achieve more robust and better performance than the traditional schemes.
Instead of improving it, in this paper, the safety of deep learning based
methods in the field of image forensics is taken into account. To the best of
our knowledge, this is a first work focusing on this topic. Specifically, we
experimentally find that the method using deep learning would fail when adding
the slight noise into the images (adversarial images). Furthermore, two kinds
of strategys are proposed to enforce security of deep learning-based method.
Firstly, an extra penalty term to the loss function is added, which is referred
to the 2-norm of the gradient of the loss with respect to the input images, and
then an novel training method are adopt to train the model by fusing the normal
and adversarial images. Experimental results show that the proposed algorithm
can achieve good performance even in the case of adversarial images and provide
a safety consideration for deep learning-based image forensic
Visual Tracking via Shallow and Deep Collaborative Model
In this paper, we propose a robust tracking method based on the collaboration
of a generative model and a discriminative classifier, where features are
learned by shallow and deep architectures, respectively. For the generative
model, we introduce a block-based incremental learning scheme, in which a local
binary mask is constructed to deal with occlusion. The similarity degrees
between the local patches and their corresponding subspace are integrated to
formulate a more accurate global appearance model. In the discriminative model,
we exploit the advances of deep learning architectures to learn generic
features which are robust to both background clutters and foreground appearance
variations. To this end, we first construct a discriminative training set from
auxiliary video sequences. A deep classification neural network is then trained
offline on this training set. Through online fine-tuning, both the hierarchical
feature extractor and the classifier can be adapted to the appearance change of
the target for effective online tracking. The collaboration of these two models
achieves a good balance in handling occlusion and target appearance change,
which are two contradictory challenging factors in visual tracking. Both
quantitative and qualitative evaluations against several state-of-the-art
algorithms on challenging image sequences demonstrate the accuracy and the
robustness of the proposed tracker.Comment: Undergraduate Thesis, appearing in Pattern Recognitio
Can Deep Learning Predict Risky Retail Investors? A Case Study in Financial Risk Behavior Forecasting
The paper examines the potential of deep learning to support decisions in
financial risk management. We develop a deep learning model for predicting
whether individual spread traders secure profits from future trades. This task
embodies typical modeling challenges faced in risk and behavior forecasting.
Conventional machine learning requires data that is representative of the
feature-target relationship and relies on the often costly development,
maintenance, and revision of handcrafted features. Consequently, modeling
highly variable, heterogeneous patterns such as trader behavior is challenging.
Deep learning promises a remedy. Learning hierarchical distributed
representations of the data in an automatic manner (e.g. risk taking behavior),
it uncovers generative features that determine the target (e.g., trader's
profitability), avoids manual feature engineering, and is more robust toward
change (e.g. dynamic market conditions). The results of employing a deep
network for operational risk forecasting confirm the feature learning
capability of deep learning, provide guidance on designing a suitable network
architecture and demonstrate the superiority of deep learning over machine
learning and rule-based benchmarks.Comment: Within the "equal" contribution, Yaodong Yang contributed the core
deep learning algorithm along with its experimental results, and the first
draft of the manuscript (including Figure 1,2,3,4,7,8,9,11, and Table 3
- …