11 research outputs found
Towards Robust Deep Neural Networks
Deep neural networks (DNNs) enable state-of-the-art performance for most machine
learning tasks. Unfortunately, they are vulnerable to attacks, such as Trojans during
training and Adversarial Examples at test time. Adversarial Examples are inputs
with carefully crafted perturbations added to benign samples. In the Computer
Vision domain, while the perturbations being imperceptible to humans, Adversarial
Examples can successfully misguide or fool DNNs. Meanwhile, Trojan or backdoor
attacks involve attackers tampering with the training process, for example, to inject
poisoned training data to embed a backdoor into the network that can be activated
during model deployment when the Trojan triggers (known only to the attackers)
appear in the model’s inputs. This dissertation investigates methods of building robust
DNNs against these training-time and test-time threats.
Recognising the threat of Adversarial Examples in the malware domain, this research
considers the problem of realising a robust DNN-based malware detector against Adversarial
Example attacks by developing a Bayesian adversarial learning algorithm. In contrast
to vision tasks, adversarial learning in a domain without a differentiable or invertible
mapping function from the problemspace (such as software code inputs) to the feature
space is hard. The study proposes an alternative; performing adversarial learning in
the feature space and proving the projection of perturbed yet, valid malware, in the
problem space into the feature space will be a subset of feature-space adversarial
attacks. The Bayesian approach improves benign performance, provably bounds
the difference between adversarial risk and empirical risk and improves robustness
against increasingly large attack budgets not employed during training.
To investigate the problem of improving the robustness of DNNs against Adversarial
Examples–carefully crafted perturbation added to inputs—in the Computer Vision
domain, the research considers the problem of developing a Bayesian learning algorithm to
realise a robust DNN against Adversarial Examples in the CV domain. Accordingly, a novel
Bayesian learning method is designed that conceptualises an information gain objective
to measure and force the information learned from both benign and Adversarial
Examples to be similar. This method proves that minimising this information gain
objective further tightens the bound of the difference between adversarial risk and empirical risk to move towards a basis for a principled method of adversarially training
BNNs.
Recognising the threat from backdoor or Trojan attacks against DNNs, the research
considers the problem of finding a robust defence method that is effective against Trojan
attacks. The research explores a new idea in the domain; sanitisation of inputs and
proposes Februus to neutralise highly potent and insidious Trojan attacks on DNN
systems at run-time. In Trojan attacks, an adversary activates a backdoor crafted in
a deep neural network model using a secret trigger, a Trojan, applied to any input
to alter the model’s decision to a target prediction—a target determined by and only
known to the attacker. Februus sanitises the incoming input by surgically removing the
potential trigger artifacts and restoring the input for the classification task. Februus
enables effective Trojan mitigation by sanitising inputs with no loss of performance
for sanitised inputs, trojaned or benign. This method is highly effective at defending
against advanced Trojan attack variants as well as challenging, adaptive attacks where
attackers have full knowledge of the defence method.
Investigating the connections between Trojan attacks and spatially constrained
Adversarial Examples or so-called Adversarial Patches in the input space, the research
exposes an emerging threat; an attack exploiting the vulnerability of a DNN to generate
naturalistic adversarial patches as universal triggers. For the first time, a method based
on Generative Adversarial Networks is developed to exploit a GAN’s latent space to
search for universal naturalistic adversarial patches. The proposed attack’s advantage
is its ability to exert a high level of control, enabling attackers to craft naturalistic
adversarial patches that are highly effective, robust against state-of-the-art DNNs, and
deployable in the physical world without needing to interfere with the model building
process or risking discovery. Until now, this has only been demonstrably possible
using Trojan attack methods.Thesis (Ph.D.) -- University of Adelaide, School of Computer Science, 202
Bayesian Learning with Information Gain Provably Bounds Risk for a Robust Adversarial Defense
We present a new algorithm to learn a deep neural network model robust
against adversarial attacks. Previous algorithms demonstrate an adversarially
trained Bayesian Neural Network (BNN) provides improved robustness. We
recognize the adversarial learning approach for approximating the multi-modal
posterior distribution of a Bayesian model can lead to mode collapse;
consequently, the model's achievements in robustness and performance are
sub-optimal. Instead, we first propose preventing mode collapse to better
approximate the multi-modal posterior distribution. Second, based on the
intuition that a robust model should ignore perturbations and only consider the
informative content of the input, we conceptualize and formulate an information
gain objective to measure and force the information learned from both benign
and adversarial training instances to be similar. Importantly. we prove and
demonstrate that minimizing the information gain objective allows the
adversarial risk to approach the conventional empirical risk. We believe our
efforts provide a step toward a basis for a principled method of adversarially
training BNNs. Our model demonstrate significantly improved robustness--up to
20%--compared with adversarial training and Adv-BNN under PGD attacks with
0.035 distortion on both CIFAR-10 and STL-10 datasets.Comment: Published at ICML 2022. Code is available at
https://github.com/baogiadoan/IG-BN
Bayesian Learned Models Can Detect Adversarial Malware For Free
The vulnerability of machine learning-based malware detectors to adversarial
attacks has prompted the need for robust solutions. Adversarial training is an
effective method but is computationally expensive to scale up to large datasets
and comes at the cost of sacrificing model performance for robustness. We
hypothesize that adversarial malware exploits the low-confidence regions of
models and can be identified using epistemic uncertainty of ML approaches --
epistemic uncertainty in a machine learning-based malware detector is a result
of a lack of similar training samples in regions of the problem space. In
particular, a Bayesian formulation can capture the model parameters'
distribution and quantify epistemic uncertainty without sacrificing model
performance. To verify our hypothesis, we consider Bayesian learning approaches
with a mutual information-based formulation to quantify uncertainty and detect
adversarial malware in Android, Windows domains and PDF malware. We found,
quantifying uncertainty through Bayesian learning methods can defend against
adversarial malware. In particular, Bayesian models: (1) are generally capable
of identifying adversarial malware in both feature and problem space, (2) can
detect concept drift by measuring uncertainty, and (3) with a
diversity-promoting approach (or better posterior approximations) lead to
parameter instances from the posterior to significantly enhance a detectors'
ability.Comment: Accepted to the 29th European Symposium on Research in Computer
Security (ESORICS) 2024 Conferenc
The global burden of cancer attributable to risk factors, 2010-19 : a systematic analysis for the Global Burden of Disease Study 2019
Background Understanding the magnitude of cancer burden attributable to potentially modifiable risk factors is crucial for development of effective prevention and mitigation strategies. We analysed results from the Global Burden of Diseases, Injuries, and Risk Factors Study (GBD) 2019 to inform cancer control planning efforts globally. Methods The GBD 2019 comparative risk assessment framework was used to estimate cancer burden attributable to behavioural, environmental and occupational, and metabolic risk factors. A total of 82 risk-outcome pairs were included on the basis of the World Cancer Research Fund criteria. Estimated cancer deaths and disability-adjusted life-years (DALYs) in 2019 and change in these measures between 2010 and 2019 are presented. Findings Globally, in 2019, the risk factors included in this analysis accounted for 4.45 million (95% uncertainty interval 4.01-4.94) deaths and 105 million (95.0-116) DALYs for both sexes combined, representing 44.4% (41.3-48.4) of all cancer deaths and 42.0% (39.1-45.6) of all DALYs. There were 2.88 million (2.60-3.18) risk-attributable cancer deaths in males (50.6% [47.8-54.1] of all male cancer deaths) and 1.58 million (1.36-1.84) risk-attributable cancer deaths in females (36.3% [32.5-41.3] of all female cancer deaths). The leading risk factors at the most detailed level globally for risk-attributable cancer deaths and DALYs in 2019 for both sexes combined were smoking, followed by alcohol use and high BMI. Risk-attributable cancer burden varied by world region and Socio-demographic Index (SDI), with smoking, unsafe sex, and alcohol use being the three leading risk factors for risk-attributable cancer DALYs in low SDI locations in 2019, whereas DALYs in high SDI locations mirrored the top three global risk factor rankings. From 2010 to 2019, global risk-attributable cancer deaths increased by 20.4% (12.6-28.4) and DALYs by 16.8% (8.8-25.0), with the greatest percentage increase in metabolic risks (34.7% [27.9-42.8] and 33.3% [25.8-42.0]). Interpretation The leading risk factors contributing to global cancer burden in 2019 were behavioural, whereas metabolic risk factors saw the largest increases between 2010 and 2019. Reducing exposure to these modifiable risk factors would decrease cancer mortality and DALY rates worldwide, and policies should be tailored appropriately to local cancer risk factor burden. Copyright (C) 2022 The Author(s). Published by Elsevier Ltd. This is an Open Access article under the CC BY 4.0 license.Peer reviewe
Safety and efficacy of fluoxetine on functional outcome after acute stroke (AFFINITY): a randomised, double-blind, placebo-controlled trial
Background
Trials of fluoxetine for recovery after stroke report conflicting results. The Assessment oF FluoxetINe In sTroke recoverY (AFFINITY) trial aimed to show if daily oral fluoxetine for 6 months after stroke improves functional outcome in an ethnically diverse population.
Methods
AFFINITY was a randomised, parallel-group, double-blind, placebo-controlled trial done in 43 hospital stroke units in Australia (n=29), New Zealand (four), and Vietnam (ten). Eligible patients were adults (aged ≥18 years) with a clinical diagnosis of acute stroke in the previous 2–15 days, brain imaging consistent with ischaemic or haemorrhagic stroke, and a persisting neurological deficit that produced a modified Rankin Scale (mRS) score of 1 or more. Patients were randomly assigned 1:1 via a web-based system using a minimisation algorithm to once daily, oral fluoxetine 20 mg capsules or matching placebo for 6 months. Patients, carers, investigators, and outcome assessors were masked to the treatment allocation. The primary outcome was functional status, measured by the mRS, at 6 months. The primary analysis was an ordinal logistic regression of the mRS at 6 months, adjusted for minimisation variables. Primary and safety analyses were done according to the patient's treatment allocation. The trial is registered with the Australian New Zealand Clinical Trials Registry, ACTRN12611000774921.
Findings
Between Jan 11, 2013, and June 30, 2019, 1280 patients were recruited in Australia (n=532), New Zealand (n=42), and Vietnam (n=706), of whom 642 were randomly assigned to fluoxetine and 638 were randomly assigned to placebo. Mean duration of trial treatment was 167 days (SD 48·1). At 6 months, mRS data were available in 624 (97%) patients in the fluoxetine group and 632 (99%) in the placebo group. The distribution of mRS categories was similar in the fluoxetine and placebo groups (adjusted common odds ratio 0·94, 95% CI 0·76–1·15; p=0·53). Compared with patients in the placebo group, patients in the fluoxetine group had more falls (20 [3%] vs seven [1%]; p=0·018), bone fractures (19 [3%] vs six [1%]; p=0·014), and epileptic seizures (ten [2%] vs two [<1%]; p=0·038) at 6 months.
Interpretation
Oral fluoxetine 20 mg daily for 6 months after acute stroke did not improve functional outcome and increased the risk of falls, bone fractures, and epileptic seizures. These results do not support the use of fluoxetine to improve functional outcome after stroke
Transferable Graph Backdoor Attack
Graph Neural Networks (GNNs) have achieved tremendous success in many graph
mining tasks benefitting from the message passing strategy that fuses the local
structure and node features for better graph representation learning. Despite
the success of GNNs, and similar to other types of deep neural networks, GNNs
are found to be vulnerable to unnoticeable perturbations on both graph
structure and node features. Many adversarial attacks have been proposed to
disclose the fragility of GNNs under different perturbation strategies to
create adversarial examples. However, vulnerability of GNNs to successful
backdoor attacks was only shown recently. In this paper, we disclose the TRAP
attack, a Transferable GRAPh backdoor attack. The core attack principle is to
poison the training dataset with perturbation-based triggers that can lead to
an effective and transferable backdoor attack. The perturbation trigger for a
graph is generated by performing the perturbation actions on the graph
structure via a gradient based score matrix from a surrogate model. Compared
with prior works, TRAP attack is different in several ways: i) it exploits a
surrogate Graph Convolutional Network (GCN) model to generate perturbation
triggers for a blackbox based backdoor attack; ii) it generates sample-specific
perturbation triggers which do not have a fixed pattern; and iii) the attack
transfers, for the first time in the context of GNNs, to different GNN models
when trained with the forged poisoned training dataset. Through extensive
evaluations on four real-world datasets, we demonstrate the effectiveness of
the TRAP attack to build transferable backdoors in four different popular GNNs
using four real-world datasets.Comment: Accepted by the 25th International Symposium on Research in Attacks,
Intrusions, and Defense
Feature-Space Bayesian Adversarial Learning Improved Malware Detector Robustness
We present a new algorithm to train a robust malware detector. Malware is a prolific problem and malware detectors are a front-line defense. Modern detectors rely on machine learning algorithms. Now, the adversarial objective is to devise alterations to the malware code to decrease the chance of being detected whilst preserving the functionality and realism of the malware. Adversarial learning is effective in improving robustness but generating functional and realistic adversarial malware samples is non-trivial. Because: i) in contrast to tasks capable of using gradient-based feedback, adversarial learning in a domain without a differentiable mapping function from the problem space (malware code inputs) to the feature space is hard; and ii) it is difficult to ensure the adversarial malware is realistic and functional.
This presents a challenge for developing scalable adversarial machine learning algorithms for large datasets at a production or commercial scale to realize robust malware detectors. We propose an alternative; perform adversarial learning in the feature space in contrast to the problem space. We prove the projection of perturbed, yet valid malware, in the problem space into feature space will always be a subset of adversarials generated in the feature space. Hence, by generating a robust network against feature-space adversarial examples, we inherently achieve robustness against problem-space adversarial examples. We formulate a Bayesian adversarial learning objective that captures the distribution of models for improved robustness.
To explain the robustness of the Bayesian adversarial learning algorithm, we prove that our learning method bounds the difference between the adversarial risk and empirical risk and improves robustness. We show that Bayesian neural networks (BNNs) achieve state-of-the-art results; especially in the False Positive Rate (FPR) regime. Adversarially trained BNNs achieve state-of-the-art robustness. Notably, adversarially trained BNNs are robust against stronger attacks with larger attack budgets by a margin of up to 15% on a recent production-scale malware dataset of more than 20 million samples. Importantly, our efforts create a benchmark for future defenses in the malware domain
Twelve-Month Outcomes of the AFFINITY Trial of Fluoxetine for Functional Recovery After Acute Stroke: AFFINITY Trial Steering Committee on Behalf of the AFFINITY Trial Collaboration
Background and Purpose: The AFFINITY trial (Assessment of Fluoxetine in Stroke Recovery) reported that oral fluoxetine 20 mg daily for 6 months after acute stroke did not improve functional outcome and increased the risk of falls, bone fractures, and seizures. After trial medication was ceased at 6 months, survivors were followed to 12 months post-randomization. This preplanned secondary analysis aimed to determine any sustained or delayed effects of fluoxetine at 12 months post-randomization. Methods: AFFINITY was a randomized, parallel-group, double-blind, placebo-controlled trial in adults (n=1280) with a clinical diagnosis of stroke in the previous 2 to 15 days and persisting neurological deficit who were recruited at 43 hospital stroke units in Australia (n=29), New Zealand (4), and Vietnam (10) between 2013 and 2019. Participants were randomized to oral fluoxetine 20 mg once daily (n=642) or matching placebo (n=638) for 6 months and followed until 12 months after randomization. The primary outcome was function, measured by the modified Rankin Scale, at 6 months. Secondary outcomes for these analyses included measures of the modified Rankin Scale, mood, cognition, overall health status, fatigue, health-related quality of life, and safety at 12 months. Results: Adherence to trial medication was for a mean 167 (SD 48) days and similar between randomized groups. At 12 months, the distribution of modified Rankin Scale categories was similar in the fluoxetine and placebo groups (adjusted common odds ratio, 0.93 [95% CI, 0.76–1.14]; P =0.46). Compared with placebo, patients allocated fluoxetine had fewer recurrent ischemic strokes (14 [2.18%] versus 29 [4.55%]; P =0.02), and no longer had significantly more falls (27 [4.21%] versus 15 [2.35%]; P =0.08), bone fractures (23 [3.58%] versus 11 [1.72%]; P =0.05), or seizures (11 [1.71%] versus 8 [1.25%]; P =0.64) at 12 months. Conclusions: Fluoxetine 20 mg daily for 6 months after acute stroke had no delayed or sustained effect on functional outcome, falls, bone fractures, or seizures at 12 months poststroke. The lower rate of recurrent ischemic stroke in the fluoxetine group is most likely a chance finding. REGISTRATION: URL: http://www.anzctr.org.au/ ; Unique identifier: ACTRN12611000774921