22 research outputs found

    Out-of-Distribution Detection Using Neural Rendering Generative Models

    Get PDF
    Out-of-distribution (OoD) detection is a natural downstream task for deep generative models, due to their ability to learn the input probability distribution. There are mainly two classes of approaches for OoD detection using deep generative models, viz., based on likelihood measure and the reconstruction loss. However, both approaches are unable to carry out OoD detection effectively, especially when the OoD samples have smaller variance than the training samples. For instance, both flow based and VAE models assign higher likelihood to images from SVHN when trained on CIFAR-10 images. We use a recently proposed generative model known as neural rendering model (NRM) and derive metrics for OoD. We show that NRM unifies both approaches since it provides a likelihood estimate and also carries out reconstruction in each layer of the neural network. Among various measures, we found the joint likelihood of latent variables to be the most effective one for OoD detection. Our results show that when trained on CIFAR-10, lower likelihood (of latent variables) is assigned to SVHN images. Additionally, we show that this metric is consistent across other OoD datasets. To the best of our knowledge, this is the first work to show consistently lower likelihood for OoD data with smaller variance with deep generative models

    MultiRobustBench: Benchmarking Robustness Against Multiple Attacks

    Full text link
    The bulk of existing research in defending against adversarial examples focuses on defending against a single (typically bounded Lp-norm) attack, but for a practical setting, machine learning (ML) models should be robust to a wide variety of attacks. In this paper, we present the first unified framework for considering multiple attacks against ML models. Our framework is able to model different levels of learner's knowledge about the test-time adversary, allowing us to model robustness against unforeseen attacks and robustness against unions of attacks. Using our framework, we present the first leaderboard, MultiRobustBench, for benchmarking multiattack evaluation which captures performance across attack types and attack strengths. We evaluate the performance of 16 defended models for robustness against a set of 9 different attack types, including Lp-based threat models, spatial transformations, and color changes, at 20 different attack strengths (180 attacks total). Additionally, we analyze the state of current defenses against multiple attacks. Our analysis shows that while existing defenses have made progress in terms of average robustness across the set of attacks used, robustness against the worst-case attack is still a big open problem as all existing models perform worse than random guessing.Comment: ICML 202

    Neural Networks with Recurrent Generative Feedback

    Get PDF
    Neural networks are vulnerable to input perturbations such as additive noise and adversarial attacks. In contrast, human perception is much more robust to such perturbations. The Bayesian brain hypothesis states that human brains use an internal generative model to update the posterior beliefs of the sensory input. This mechanism can be interpreted as a form of self-consistency between the maximum a posteriori (MAP) estimation of an internal generative model and the external environment. Inspired by such hypothesis, we enforce self-consistency in neural networks by incorporating generative recurrent feedback. We instantiate this design on convolutional neural networks (CNNs). The proposed framework, termed Convolutional Neural Networks with Feedback (CNN-F), introduces a generative feedback with latent variables to existing CNN architectures, where consistent predictions are made through alternating MAP inference under a Bayesian framework. In the experiments, CNN-F shows considerably improved adversarial robustness over conventional feedforward CNNs on standard benchmarks

    Neural Networks with Recurrent Generative Feedback

    Get PDF
    Neural networks are vulnerable to input perturbations such as additive noise and adversarial attacks. In contrast, human perception is much more robust to such perturbations. The Bayesian brain hypothesis states that human brains use an internal generative model to update the posterior beliefs of the sensory input. This mechanism can be interpreted as a form of self-consistency between the maximum a posteriori (MAP) estimation of an internal generative model and the external environment. Inspired by such hypothesis, we enforce self-consistency in neural networks by incorporating generative recurrent feedback. We instantiate this design on convolutional neural networks (CNNs). The proposed framework, termed Convolutional Neural Networks with Feedback (CNN-F), introduces a generative feedback with latent variables to existing CNN architectures, where consistent predictions are made through alternating MAP inference under a Bayesian framework. In the experiments, CNN-F shows considerably improved adversarial robustness over conventional feedforward CNNs on standard benchmarks.Comment: NeurIPS 202

    Characterizing the Optimal 0-1 Loss for Multi-class Classification with a Test-time Attacker

    Full text link
    Finding classifiers robust to adversarial examples is critical for their safe deployment. Determining the robustness of the best possible classifier under a given threat model for a given data distribution and comparing it to that achieved by state-of-the-art training methods is thus an important diagnostic tool. In this paper, we find achievable information-theoretic lower bounds on loss in the presence of a test-time attacker for multi-class classifiers on any discrete dataset. We provide a general framework for finding the optimal 0-1 loss that revolves around the construction of a conflict hypergraph from the data and adversarial constraints. We further define other variants of the attacker-classifier game that determine the range of the optimal loss more efficiently than the full-fledged hypergraph construction. Our evaluation shows, for the first time, an analysis of the gap to optimal robustness for classifiers in the multi-class setting on benchmark datasets.Comment: NeurIPS 2023 Spotligh

    Out-of-Distribution Detection Using Neural Rendering Generative Models

    Get PDF
    Out-of-distribution (OoD) detection is a natural downstream task for deep generative models, due to their ability to learn the input probability distribution. There are mainly two classes of approaches for OoD detection using deep generative models, viz., based on likelihood measure and the reconstruction loss. However, both approaches are unable to carry out OoD detection effectively, especially when the OoD samples have smaller variance than the training samples. For instance, both flow based and VAE models assign higher likelihood to images from SVHN when trained on CIFAR-10 images. We use a recently proposed generative model known as neural rendering model (NRM) and derive metrics for OoD. We show that NRM unifies both approaches since it provides a likelihood estimate and also carries out reconstruction in each layer of the neural network. Among various measures, we found the joint likelihood of latent variables to be the most effective one for OoD detection. Our results show that when trained on CIFAR-10, lower likelihood (of latent variables) is assigned to SVHN images. Additionally, we show that this metric is consistent across other OoD datasets. To the best of our knowledge, this is the first work to show consistently lower likelihood for OoD data with smaller variance with deep generative models

    Formulating Robustness Against Unforeseen Attacks

    Full text link
    Existing defenses against adversarial examples such as adversarial training typically assume that the adversary will conform to a specific or known threat model, such as â„“p\ell_p perturbations within a fixed budget. In this paper, we focus on the scenario where there is a mismatch in the threat model assumed by the defense during training, and the actual capabilities of the adversary at test time. We ask the question: if the learner trains against a specific "source" threat model, when can we expect robustness to generalize to a stronger unknown "target" threat model during test-time? Our key contribution is to formally define the problem of learning and generalization with an unforeseen adversary, which helps us reason about the increase in adversarial risk from the conventional perspective of a known adversary. Applying our framework, we derive a generalization bound which relates the generalization gap between source and target threat models to variation of the feature extractor, which measures the expected maximum difference between extracted features across a given threat model. Based on our generalization bound, we propose adversarial training with variation regularization (AT-VR) which reduces variation of the feature extractor across the source threat model during training. We empirically demonstrate that AT-VR can lead to improved generalization to unforeseen attacks during test-time compared to standard adversarial training on Gaussian and image datasets

    Automatic Brix Measurement for Watermelon Breeding

    No full text
    Sweetness or sugar content, represented by soluble solids contents (SSC), is a vital quality trait in watermelon breeding which can be assessed by the refractive index method. However, sampling watermelon juice out of the pulp is a process that is both labor-intensive and error-prone. In this study, we developed an automatic SSC measurement system for watermelon breeding to improve efficiency and decrease costs. First, we built an automatic cutting system to cut watermelons into precise halves, in which a laser rangefinder is used to measure the distance from the upper surface of the watermelon to itself, and thus, the diameter is estimated. The experiments showed a high correlation between the estimated diameters and the ground truths, with and . Then, we built an automatic Brix measurement system to obtain the Brix data from a central point on the watermelon’s section, where an image analysis procedure is applied to locate the testing point. This is then transformed to the camera coordination system, and a refractometer is driven by a 3-axis robotic arm to reach the testing point. Brix measurement experiments were conducted using three vertical gaps and four lateral gaps between the probe of the refractometer and the pulp. The result showed that the best parameters were a vertical gap of 4 mm and a lateral gap of 2 mm. The average accuracy reached 98.74%, which indicates that this study has the potential to support watermelon breeding research

    Automatic Brix Measurement for Watermelon Breeding

    No full text
    Sweetness or sugar content, represented by soluble solids contents (SSC), is a vital quality trait in watermelon breeding which can be assessed by the refractive index method. However, sampling watermelon juice out of the pulp is a process that is both labor-intensive and error-prone. In this study, we developed an automatic SSC measurement system for watermelon breeding to improve efficiency and decrease costs. First, we built an automatic cutting system to cut watermelons into precise halves, in which a laser rangefinder is used to measure the distance from the upper surface of the watermelon to itself, and thus, the diameter is estimated. The experiments showed a high correlation between the estimated diameters and the ground truths, with and . Then, we built an automatic Brix measurement system to obtain the Brix data from a central point on the watermelon’s section, where an image analysis procedure is applied to locate the testing point. This is then transformed to the camera coordination system, and a refractometer is driven by a 3-axis robotic arm to reach the testing point. Brix measurement experiments were conducted using three vertical gaps and four lateral gaps between the probe of the refractometer and the pulp. The result showed that the best parameters were a vertical gap of 4 mm and a lateral gap of 2 mm. The average accuracy reached 98.74%, which indicates that this study has the potential to support watermelon breeding research
    corecore