30 research outputs found

    Reachability Analysis for Neural Feedback Systems Using Regressive Polynomial Rule Inference

    Get PDF
    We present an approach to construct reachable set overapproxi- mations for continuous-time dynamical systems controlled using neural network feedback systems. Feedforward deep neural net- works are now widely used as a means for learning control laws through techniques such as reinforcement learning and data-driven predictive control. However, the learning algorithms for these net- works do not guarantee correctness properties on the resulting closed-loop systems. Our approach seeks to construct overapproxi- mate reachable sets by integrating a Taylor model-based flowpipe construction scheme for continuous differential equations with an approach that replaces the neural network feedback law for a small subset of inputs by a polynomial mapping. We generate the polynomial mapping using regression from input-output sam- ples. To ensure soundness, we rigorously quantify the gap between the output of the network and that of the polynomial model. We demonstrate the effectiveness of our approach over a suite of bench- mark examples ranging from 2 to 17 state variables, comparing our approach with alternative ideas based on range analysis

    Guaranteed Conformance of Neurosymbolic Models to Natural Constraints

    Get PDF
    Deep neural networks have emerged as the workhorse for a large section of robotics and control applications, especially as models for dynamical systems. Such data-driven models are in turn used for designing and verifying autonomous systems. This is particularly useful in modeling medical systems where data can be leveraged to individualize treatment. In safety-critical applications, it is important that the data-driven model is conformant to established knowledge from the natural sciences. Such knowledge is often available or can often be distilled into a (possibly black-box) model MM. For instance, the unicycle model for an F1 racing car. In this light, we consider the following problem - given a model MM and state transition dataset, we wish to best approximate the system model while being bounded distance away from MM. We propose a method to guarantee this conformance. Our first step is to distill the dataset into few representative samples called memories, using the idea of a growing neural gas. Next, using these memories we partition the state space into disjoint subsets and compute bounds that should be respected by the neural network, when the input is drawn from a particular subset. This serves as a symbolic wrapper for guaranteed conformance. We argue theoretically that this only leads to bounded increase in approximation error; which can be controlled by increasing the number of memories. We experimentally show that on three case studies (Car Model, Drones, and Artificial Pancreas), our constrained neurosymbolic models conform to specified MM models (each encoding various constraints) with order-of-magnitude improvements compared to the augmented Lagrangian and vanilla training methods

    Guaranteed Conformance of Neurosymbolic Models to Natural Constraints

    Get PDF
    Deep neural networks have emerged as the workhorse for a large section of robotics and control applications, especially as models for dynamical systems. Such data-driven models are in turn used for designing and verifying autonomous systems. This is particularly useful in modeling medical systems where data can be leveraged to individualize treatment. In safety-critical applications, it is important that the data-driven model is conformant to established knowledge from the natural sciences. Such knowledge is often available or can often be distilled into a (possibly black-box) model M. For instance, the unicycle model for an F1 racing car. In this light, we consider the following problem - given a model M and state transition dataset, we wish to best approximate the system model while being bounded distance away from M. We propose a method to guarantee this conformance. Our first step is to distill the dataset into few representative samples called memories, using the idea of a growing neural gas. Next, using these memories we partition the state space into disjoint subsets and compute bounds that should be respected by the neural network, when the input is drawn from a particular subset. This serves as a symbolic wrapper for guaranteed conformance. We argue theoretically that this only leads to bounded increase in approximation error; which can be controlled by increasing the number of memories. We experimentally show that on three case studies (Car Model, Drones, and Artificial Pancreas), our constrained neurosymbolic models conform to specified M models (each encoding various constraints) with order-of-magnitude improvements compared to the augmented Lagrangian and vanilla training methods

    Outcomes from Early Experience with Laparoscopic Inguinal Hernia Repair Versus Open Technique: Navigating the Learning Curve

    Get PDF
    Objectives: The current consensus in literature often suggests laparoscopic inguinal hernia repair (LIHR) as superior to open inguinal hernia repair (OIHR) regarding postoperative pain, recurrence rates, duration of hospital stay, and other postoperative outcomes. Our study aimed to evaluate these outcomes within the context of our centre in its initial experience of laparoscopic repairs. Methods: We performed a single-centre, retrospective observational study encompassing all patients who underwent elective OIHR and LIHR from January 2011 through September 2020. This comprised 2690 and 158 cases respectively. examining parameters like demographic data, comorbidities, hernia type, mesh characteristics, surgery duration, hospital stay, and immediate postoperative complications. Results: The demographic profiles, hospital stay, and complication rates were similar in both groups. However, surgical site infection was present exclusively in the OIHR group (3.5% vs. 0.0%; p<0.05). The timeline for returning to normal activities was statistically shorter for the LIHR group [6 days vs. 8 days; p <0.05]. The most frequent immediate complication in the LIHR group was subcutaneous emphysema [46.67%; p<0.05]. Recurrence [9.23% vs. 3.6%; p=0.09] and chronic pain [41.5% vs. 13.6%; p<0.05] were higher in the LIHR group. Conclusion: In the course of our early experience with LIHR, we observed lower recurrence and chronic pain rates with OIHR. However, LIHR had significant advantages with respect to faster patient recovery and lower rates of SSI. While our results contribute an interesting deviation from the standard narrative, they should be interpreted within the context of a learning curve associated with our early experience with LIHR. Keywords: Hernia; Hernia, Inguinal; Laparoscopy

    Memory-Consistent Neural Networks for Imitation Learning

    Full text link
    Imitation learning considerably simplifies policy synthesis compared to alternative approaches by exploiting access to expert demonstrations. For such imitation policies, errors away from the training samples are particularly critical. Even rare slip-ups in the policy action outputs can compound quickly over time, since they lead to unfamiliar future states where the policy is still more likely to err, eventually causing task failures. We revisit simple supervised ``behavior cloning'' for conveniently training the policy from nothing more than pre-recorded demonstrations, but carefully design the model class to counter the compounding error phenomenon. Our ``memory-consistent neural network'' (MCNN) outputs are hard-constrained to stay within clearly specified permissible regions anchored to prototypical ``memory'' training samples. We provide a guaranteed upper bound for the sub-optimality gap induced by MCNN policies. Using MCNNs on 9 imitation learning tasks, with MLP, Transformer, and Diffusion backbones, spanning dexterous robotic manipulation and driving, proprioceptive inputs and visual inputs, and varying sizes and types of demonstration data, we find large and consistent gains in performance, validating that MCNNs are better-suited than vanilla deep neural networks for imitation learning applications. Website: https://sites.google.com/view/mcnn-imitationComment: 22 pages (9 main pages

    Credal Bayesian Deep Learning

    Get PDF
    Uncertainty quantification and robustness to distribution shifts are important goals in machine learning and artificial intelligence. Although Bayesian Neural Networks (BNNs) allow for uncertainty in the predictions to be assessed, different sources of uncertainty are indistinguishable. We present Credal Bayesian Deep Learning (CBDL). Heuristically, CBDL allows to train an (uncountably) infinite ensemble of BNNs, using only finitely many elements. This is possible thanks to prior and likelihood finitely generated credal sets (FGCSs), a concept from the imprecise probability literature. Intuitively, convex combinations of a finite collection of prior-likelihood pairs are able to represent infinitely many such pairs. After training, CBDL outputs a set of posteriors on the parameters of the neural network. At inference time, such posterior set is used to derive a set of predictive distributions that is in turn utilized to distinguish between aleatoric and epistemic uncertainties, and to quantify them. The predictive set also produces either (i) a collection of outputs enjoying desirable probabilistic guarantees, or (ii) the single output that is deemed the best, that is, the one having the highest predictive lower probability -- another imprecise-probabilistic concept. CBDL is more robust than single BNNs to prior and likelihood misspecification, and to distribution shift. We show that CBDL is better at quantifying and disentangling different types of uncertainties than single BNNs and ensemble of BNNs.In addition, we apply CBDL to two case studies to demonstrate its downstream tasks capabilities: one, for motion prediction in autonomous driving scenarios, and two, to model blood glucose and insulin dynamics for artificial pancreas control. We show that CBDL performs better when compared to an ensemble of BNNs baseline

    Exploring with Sticky Mittens: Reinforcement Learning with Expert Interventions via Option Templates

    Get PDF
    Long horizon robot learning tasks with sparse rewards pose a significant challenge for current reinforcement learning algorithms. A key feature enabling humans to learn challenging control tasks is that they often receive expert intervention that enables them to understand the high-level structure of the task before mastering low-level control actions. We propose a framework for leveraging expert intervention to solve long-horizon reinforcement learning tasks. We consider option templates, which are specifications encoding a potential option that can be trained using reinforcement learning. We formulate expert intervention as allowing the agent to execute option templates before learning an implementation. This enables them to use an option, before committing costly resources to learning it. We evaluate our approach on three challenging reinforcement learning problems, showing that it outperforms state-of-the-art approaches by two orders of magnitude

    Distributionally Robust Statistical Verification with Imprecise Neural Networks

    Full text link
    A particularly challenging problem in AI safety is providing guarantees on the behavior of high-dimensional autonomous systems. Verification approaches centered around reachability analysis fail to scale, and purely statistical approaches are constrained by the distributional assumptions about the sampling process. Instead, we pose a distributionally robust version of the statistical verification problem for black-box systems, where our performance guarantees hold over a large family of distributions. This paper proposes a novel approach based on a combination of active learning, uncertainty quantification, and neural network verification. A central piece of our approach is an ensemble technique called Imprecise Neural Networks, which provides the uncertainty to guide active learning. The active learning uses an exhaustive neural-network verification tool Sherlock to collect samples. An evaluation on multiple physical simulators in the openAI gym Mujoco environments with reinforcement-learned controllers demonstrates that our approach can provide useful and scalable guarantees for high-dimensional systems
    corecore