9 research outputs found

    PGN: A perturbation generation network against deep reinforcement learning

    Full text link
    Deep reinforcement learning has advanced greatly and applied in many areas. In this paper, we explore the vulnerability of deep reinforcement learning by proposing a novel generative model for creating effective adversarial examples to attack the agent. Our proposed model can achieve both targeted attacks and untargeted attacks. Considering the specificity of deep reinforcement learning, we propose the action consistency ratio as a measure of stealthiness, and a new measurement index of effectiveness and stealthiness. Experiment results show that our method can ensure the effectiveness and stealthiness of attack compared with other algorithms. Moreover, our methods are considerably faster and thus can achieve rapid and efficient verification of the vulnerability of deep reinforcement learning

    Detecting Adversarial Directions in Deep Reinforcement Learning to Make Robust Decisions

    Full text link
    Learning in MDPs with highly complex state representations is currently possible due to multiple advancements in reinforcement learning algorithm design. However, this incline in complexity, and furthermore the increase in the dimensions of the observation came at the cost of volatility that can be taken advantage of via adversarial attacks (i.e. moving along worst-case directions in the observation space). To solve this policy instability problem we propose a novel method to detect the presence of these non-robust directions via local quadratic approximation of the deep neural policy loss. Our method provides a theoretical basis for the fundamental cut-off between safe observations and adversarial observations. Furthermore, our technique is computationally efficient, and does not depend on the methods used to produce the worst-case directions. We conduct extensive experiments in the Arcade Learning Environment with several different adversarial attack techniques. Most significantly, we demonstrate the effectiveness of our approach even in the setting where non-robust directions are explicitly optimized to circumvent our proposed method.Comment: Published in ICML 202

    Provable Adversarial Robustness for Group Equivariant Tasks: Graphs, Point Clouds, Molecules, and More

    Full text link
    A machine learning model is traditionally considered robust if its prediction remains (almost) constant under input perturbations with small norm. However, real-world tasks like molecular property prediction or point cloud segmentation have inherent equivariances, such as rotation or permutation equivariance. In such tasks, even perturbations with large norm do not necessarily change an input's semantic content. Furthermore, there are perturbations for which a model's prediction explicitly needs to change. For the first time, we propose a sound notion of adversarial robustness that accounts for task equivariance. We then demonstrate that provable robustness can be achieved by (1) choosing a model that matches the task's equivariances (2) certifying traditional adversarial robustness. Certification methods are, however, unavailable for many models, such as those with continuous equivariances. We close this gap by developing the framework of equivariance-preserving randomized smoothing, which enables architecture-agnostic certification. We additionally derive the first architecture-specific graph edit distance certificates, i.e. sound robustness guarantees for isomorphism equivariant tasks like node classification. Overall, a sound notion of robustness is an important prerequisite for future work at the intersection of robust and geometric machine learning.Comment: Accepted at NeurIPS 202

    Republican freedom and the rule of law

    Get PDF

    Integrating mātauranga Māori into community resilience frameworks for the built environment

    Get PDF
    Efforts to improve community resilience have seen universal efforts to grow capabilities in disaster management for the built environment. Although comprehensive frameworks exist, the perspective used to derive these frameworks fails to address marginalised communities' vulnerabilities, including the Indigenous Māori people. Colonisation and socio-economic deprivation of Māori mean community response frameworks fail to reflect their values and specific needs when preparing and responding to natural disasters. The nature of Māori culture and values presents an opportunity to reimagine the scope of community resilience for an inclusive framework. The Māori history has resilient practices weaved into the very fabric of their culture through the oral transmission of waiata, whakataukī, whakairo, and pūrākau. The practice of community resilience amongst ancient Māori was not consigned to singular events but is a holistic perspective bound to their manner of living. Sourcing and qualifying this information from people aligned and immersed in nature is the knowledge required to improve the frameworks to respond to natural disasters while expanding the literature on community resilience. This study collaborated with Ngāti Toa to represent the Māori worldview using tikanga Māori and Kaupapa Māori methodology in an action-based participatory research approach. The research analyzed traditional Māori resilience through wānanga and korero kanohi-ki-te-kanohi and evaluated various mātauranga, tikanga, and kaupapa that relate to the resilient nature of Ngāti Toa. Modern Māori perspectives of community resilience were compared to this information using a realistic earthquake scenario in the Wellington region. This led to expanding community resilience to be more inclusive of Māori communities and informed measures to enhance their resilience and address their unique needs following a disaster. This study found that the Ngāti Toa community employs traditional resilience practices dating back to their migration to Aotearoa from Hawaiki. These practices are community-led and grounded in unity, with resilience being viewed as a collective responsibility. The study identified various Te Ao Māori resilience principles by translating traditional resilience principles into corresponding Te Ao Māori principles. A response and recovery plan was developed for Ngāti Toa in response to a realistic earthquake scenario. This plan includes a vulnerability matrix and a community resilience calculator to inform Ngāti Toa of their vulnerability level and needs following a disaster. The community resilience frameworks developed for the Wellington region in response to the 7.5 magnitude earthquake were found to be economically dependent and did not utilize the response capabilities and strengths of Ngāti Toa. The frameworks lacked proper representation from Ngāti Toa and failed to address the needs of the Māori community in Wellington, resulting in inequitable outcomes. To rectify this, Page | iii community resilience frameworks for Māori should adopt a principle-based approach that supports collaborative engagement and integrates Māori cultural values. Unity is a crucial strength underpinning Māori resilience. Technological mediums such as retrofitting Marae should be developed to enhance Māori response capabilities and harness the strengths of Te Ao Māori resilience

    RESCUE: Evaluation of a Fragmented Secret Share System in Distributed-Cloud Architecture

    Get PDF
    Scaling big data infrastructure using multi-cloud environment has led to the demand for highly secure, resilient and reliable data sharing method. Several variants of secret sharing scheme have been proposed but there remains a gap in knowledge on the evaluation of these methods in relation to scalability, resilience and key management as volume of files generated increase and cloud outages persist. In line with these, this thesis presents an evaluation of a method that combines data fragmentation with Shamir’s secret sharing scheme known as Fragmented Secret Share System (FSSS). It applies data fragmentation using a calculated optimum fragment size and encrypts each fragment using a 256-bit AES key length before dispersal to cloudlets, the encryption key is managed using secret sharing methods as used in cryptography.Four experiments were performed to measure the scalability, resilience and reliability in key management. The first and second experiments evaluated scalability using defined fragment blocks and an optimum fragment size. These fragment types were used to break file of varied sizes into fragments, and then encrypted and dispersed to the cloud, and recovered when required. Both were used in combination of different secret sharing policies for key management. The third experiment tested file recovery during cloud failures, while the fourth experiment focused on efficient key management.The contributions of this thesis are of two ways: development of evaluation frameworks to measure scalability and resilience of data sharing methods; and the provision of information on relationships between file sizes and share policies combinations. While the first aimed at providing platform to measure scalability from the point of continuous production as file size and volume increase, and resilience as the potential to continue operation despite cloud outages; the second provides experimental frameworks on the effects of file sizes and share policies on overall system performance.The results of evaluation of FSSS with similar methods showed that the fragmentation method has less overhead costs irrespective of file sizes and the share policy combination. That the inherent challenges in secret sharing scheme can only be solved through alternative means such as combining secret sharing with other data fragmentation method. In all, the system is less of any erasure coding technique, making it difficult to detect corrupt or lost fragment during file recovery

    Adversarial Robust Deep Reinforcement Learning Requires Redefining Robustness

    No full text
    Learning from raw high dimensional data via interaction with a given environment has been effectively achieved through the utilization of deep neural networks. Yet the observed degradation in policy performance caused by imperceptible worst-case policy dependent translations along high sensitivity directions (i.e. adversarial perturbations) raises concerns on the robustness of deep reinforcement learning policies. In our paper, we show that these high sensitivity directions do not lie only along particular worst-case directions, but rather are more abundant in the deep neural policy landscape and can be found via more natural means in a black-box setting. Furthermore, we show that vanilla training techniques intriguingly result in learning more robust policies compared to the policies learnt via the state-of-the-art adversarial training techniques. We believe our work lays out intriguing properties of the deep reinforcement learning policy manifold and our results can help to build robust and generalizable deep reinforcement learning policies
    corecore