66 research outputs found

    Imbalanced Gradients: A Subtle Cause of Overestimated Adversarial Robustness

    Full text link
    Evaluating the robustness of a defense model is a challenging task in adversarial robustness research. Obfuscated gradients, a type of gradient masking, have previously been found to exist in many defense methods and cause a false signal of robustness. In this paper, we identify a more subtle situation called Imbalanced Gradients that can also cause overestimated adversarial robustness. The phenomenon of imbalanced gradients occurs when the gradient of one term of the margin loss dominates and pushes the attack towards to a suboptimal direction. To exploit imbalanced gradients, we formulate a Margin Decomposition (MD) attack that decomposes a margin loss into individual terms and then explores the attackability of these terms separately via a two-stage process. We also propose a MultiTargeted and an ensemble version of our MD attack. By investigating 17 defense models proposed since 2018, we find that 6 models are susceptible to imbalanced gradients and our MD attack can decrease their robustness evaluated by the best baseline standalone attack by another 2%. We also provide an in-depth analysis of the likely causes of imbalanced gradients and effective countermeasures.Comment: 19 pages, 7 figue

    Eureka: Human-Level Reward Design via Coding Large Language Models

    Full text link
    Large Language Models (LLMs) have excelled as high-level semantic planners for sequential decision-making tasks. However, harnessing them to learn complex low-level manipulation tasks, such as dexterous pen spinning, remains an open problem. We bridge this fundamental gap and present Eureka, a human-level reward design algorithm powered by LLMs. Eureka exploits the remarkable zero-shot generation, code-writing, and in-context improvement capabilities of state-of-the-art LLMs, such as GPT-4, to perform evolutionary optimization over reward code. The resulting rewards can then be used to acquire complex skills via reinforcement learning. Without any task-specific prompting or pre-defined reward templates, Eureka generates reward functions that outperform expert human-engineered rewards. In a diverse suite of 29 open-source RL environments that include 10 distinct robot morphologies, Eureka outperforms human experts on 83% of the tasks, leading to an average normalized improvement of 52%. The generality of Eureka also enables a new gradient-free in-context learning approach to reinforcement learning from human feedback (RLHF), readily incorporating human inputs to improve the quality and the safety of the generated rewards without model updating. Finally, using Eureka rewards in a curriculum learning setting, we demonstrate for the first time, a simulated Shadow Hand capable of performing pen spinning tricks, adeptly manipulating a pen in circles at rapid speed.Comment: Project website and open-source code: https://eureka-research.github.io

    Re-ViLM: Retrieval-Augmented Visual Language Model for Zero and Few-Shot Image Captioning

    Full text link
    Augmenting pretrained language models (LMs) with a vision encoder (e.g., Flamingo) has obtained the state-of-the-art results in image-to-text generation. However, these models store all the knowledge within their parameters, thus often requiring enormous model parameters to model the abundant visual concepts and very rich textual descriptions. Additionally, they are inefficient in incorporating new data, requiring a computational-expensive fine-tuning process. In this work, we introduce a Retrieval-augmented Visual Language Model, Re-ViLM, built upon the Flamingo, that supports retrieving the relevant knowledge from the external database for zero and in-context few-shot image-to-text generations. By storing certain knowledge explicitly in the external database, our approach reduces the number of model parameters and can easily accommodate new data during evaluation by simply updating the database. We also construct an interleaved image and text data that facilitates in-context few-shot learning capabilities. We demonstrate that Re-ViLM significantly boosts performance for image-to-text generation tasks, especially for zero-shot and few-shot generation in out-of-domain settings with 4 times less parameters compared with baseline methods.Comment: Findings of EMNLP 202

    Overexpressed transferrin receptor implied poor prognosis and relapse in gastrointestinal stromal tumors

    Get PDF
    Ferroptosis, as a novel-induced programmed cell death, plays critical roles in the pathogenesis of cancers. However, the promising biomarkers of ferroptosis in gastrointestinal stromal tumor (GIST) remain to be elucidated. Herein, the expression of ferroptosis-related genes was analyzed in GIST. Among the 64 ferroptosis-related genes, transferrin receptor (TFRC) expression presented a remarkable upregulation in high-risk patients through Gene Expression Omnibus (GEO) dataset analysis, as well as its significant change after imatinib was treated. Kyoto Encyclopedia of Genes and Genomes (KEGG) pathway enrichment analysis of TFRC-relevant genes revealed that TFRC expression was closely associated with cell growth pathways and metabolism-related pathways. Furthermore, patients at high risk of recurrence were more likely to exhibit high TFRC expression by immunohistochemistry. Additionally, high TFRC expression indicated an undesirable state of patient relapse, which could serve as a powerful significant independent predictor of recurrence-free survival (RFS). In summary, we systematically summarize the expression characteristics and clinical relevance of TFRC and show that TFRC can be used as a prognostic factor, which can be considered a potential therapeutic target in GIST

    Epac1 mediates protein kinase A–independent mechanism of forskolin-activated intestinal chloride secretion

    Get PDF
    Intestinal Cl− secretion is stimulated by cyclic AMP (cAMP) and intracellular calcium ([Ca2+]i). Recent studies show that protein kinase A (PKA) and the exchange protein directly activated by cAMP (Epac) are downstream targets of cAMP. Therefore, we tested whether both PKA and Epac are involved in forskolin (FSK)/cAMP-stimulated Cl− secretion. Human intestinal T84 cells and mouse small intestine were used for short circuit current (Isc) measurement in response to agonist-stimulated Cl− secretion. FSK-stimulated Cl− secretion was completely inhibited by the additive effects of the PKA inhibitor, H89 (1 µM), and the [Ca2+]i chelator, 1,2-bis-(o-aminophenoxy)-ethane-N,N,N’,N’-tetraacetic acid, tetraacetoxymethyl ester (BAPTA-AM; 25 µM). Both FSK and the Epac activator 8-pCPT-2’-O-Me-cAMP (50 µM) elevated [Ca2+]i, activated Ras-related protein 2, and induced Cl− secretion in intact or basolateral membrane–permeabilized T84 cells and mouse ileal sheets. The effects of 8-pCPT-2’-O-Me-cAMP were completely abolished by BAPTA-AM, but not by H89. In contrast, T84 cells with silenced Epac1 had a reduced Isc response to FSK, and this response was completely inhibited by H89, but not by the phospholipase C inhibitor U73122 or BAPTA-AM. The stimulatory effect of 8-pCPT-2’-O-Me-cAMP on Cl− secretion was not abolished by cystic fibrosis transmembrane conductance (CFTR) inhibitor 172 or glibenclamide, suggesting that CFTR channels are not involved. This was confirmed by lack of effect of 8-pCPT-2’-O-Me-cAMP on whole cell patch clamp recordings of CFTR currents in Chinese hamster ovary cells transiently expressing the human CFTR channel. Furthermore, biophysical characterization of the Epac1-dependent Cl− conductance of T84 cells mounted in Ussing chambers suggested that this conductance was hyperpolarization activated, inwardly rectifying, and displayed a Cl−>Br−>I− permeability sequence. These results led us to conclude that the Epac-Rap-PLC-[Ca2+]i signaling pathway is involved in cAMP-stimulated Cl− secretion, which is carried by a novel, previously undescribed Cl− channel

    Open X-Embodiment:Robotic learning datasets and RT-X models

    Get PDF
    Large, high-capacity models trained on diverse datasets have shown remarkable successes on efficiently tackling downstream applications. In domains from NLP to Computer Vision, this has led to a consolidation of pretrained models, with general pretrained backbones serving as a starting point for many applications. Can such a consolidation happen in robotics? Conventionally, robotic learning methods train a separate model for every application, every robot, and even every environment. Can we instead train "generalist" X-robot policy that can be adapted efficiently to new robots, tasks, and environments? In this paper, we provide datasets in standardized data formats and models to make it possible to explore this possibility in the context of robotic manipulation, alongside experimental results that provide an example of effective X-robot policies. We assemble a dataset from 22 different robots collected through a collaboration between 21 institutions, demonstrating 527 skills (160266 tasks). We show that a high-capacity model trained on this data, which we call RT-X, exhibits positive transfer and improves the capabilities of multiple robots by leveraging experience from other platforms. The project website is robotics-transformer-x.github.io
    • …
    corecore