142 research outputs found

    LipsFormer: Introducing Lipschitz Continuity to Vision Transformers

    Full text link
    We present a Lipschitz continuous Transformer, called LipsFormer, to pursue training stability both theoretically and empirically for Transformer-based models. In contrast to previous practical tricks that address training instability by learning rate warmup, layer normalization, attention formulation, and weight initialization, we show that Lipschitz continuity is a more essential property to ensure training stability. In LipsFormer, we replace unstable Transformer component modules with Lipschitz continuous counterparts: CenterNorm instead of LayerNorm, spectral initialization instead of Xavier initialization, scaled cosine similarity attention instead of dot-product attention, and weighted residual shortcut. We prove that these introduced modules are Lipschitz continuous and derive an upper bound on the Lipschitz constant of LipsFormer. Our experiments show that LipsFormer allows stable training of deep Transformer architectures without the need of careful learning rate tuning such as warmup, yielding a faster convergence and better generalization. As a result, on the ImageNet 1K dataset, LipsFormer-Swin-Tiny based on Swin Transformer training for 300 epochs can obtain 82.7\% without any learning rate warmup. Moreover, LipsFormer-CSwin-Tiny, based on CSwin, training for 300 epochs achieves a top-1 accuracy of 83.5\% with 4.7G FLOPs and 24M parameters. The code will be released at \url{https://github.com/IDEA-Research/LipsFormer}.Comment: To appear in ICLR 2023, our code will be public at https://github.com/IDEA-Research/LipsForme

    Fight Fire with Fire: Combating Adversarial Patch Attacks using Pattern-randomized Defensive Patches

    Full text link
    Object detection has found extensive applications in various tasks, but it is also susceptible to adversarial patch attacks. Existing defense methods often necessitate modifications to the target model or result in unacceptable time overhead. In this paper, we adopt a counterattack approach, following the principle of "fight fire with fire," and propose a novel and general methodology for defending adversarial attacks. We utilize an active defense strategy by injecting two types of defensive patches, canary and woodpecker, into the input to proactively probe or weaken potential adversarial patches without altering the target model. Moreover, inspired by randomization techniques employed in software security, we employ randomized canary and woodpecker injection patterns to defend against defense-aware attacks. The effectiveness and practicality of the proposed method are demonstrated through comprehensive experiments. The results illustrate that canary and woodpecker achieve high performance, even when confronted with unknown attack methods, while incurring limited time overhead. Furthermore, our method also exhibits sufficient robustness against defense-aware attacks, as evidenced by adaptive attack experiments

    Extracellular vesicles for breast cancer diagnosis and therapy

    Get PDF
    Breast cancer is still suffering from its poor diagnosis and the lack of effective treatment. Despite of recent development of some novel chemicals, which are found to have inspiring therapeutic effects in vitro, their outcomes in clinic trails are disappointing, mainly due to the lack of suitable therapeutic vehicles. Thanks to their ability to encapsulate bio-molecules, extracellular vesicles (EVs), including exosomes, microvesicles, and apoptotic bodies, hold great promise in becoming a suitable candidate in the breast cancer diagnosis and therapy. Currently, EVs are increasingly evaluated as potential indicators in the diagnosis of breast cancer since they are actively involved in different stages of breast cancer development, including promoting cancer occurrence and metastasis, establishing tumor ecology, and promoting tumor growth. Moreover, they are also considered as promising new platforms in breast cancer therapy. Here, we discuss the potential therapeutic applications of EVs, including EVs as biomarkers for diagnosis and therapeutic drug delivery to tumor sites. The promising data and technologies indicate the potential applicability of EVs in clinical management of patients with breast cancer.publishedVersio

    CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds

    Full text link
    We present a novel two-stage fully sparse convolutional 3D object detection framework, named CAGroup3D. Our proposed method first generates some high-quality 3D proposals by leveraging the class-aware local group strategy on the object surface voxels with the same semantic predictions, which considers semantic consistency and diverse locality abandoned in previous bottom-up approaches. Then, to recover the features of missed voxels due to incorrect voxel-wise segmentation, we build a fully sparse convolutional RoI pooling module to directly aggregate fine-grained spatial information from backbone for further proposal refinement. It is memory-and-computation efficient and can better encode the geometry-specific features of each 3D proposal. Our model achieves state-of-the-art 3D detection performance with remarkable gains of +\textit{3.6\%} on ScanNet V2 and +\textit{2.6}\% on SUN RGB-D in term of [email protected]. Code will be available at https://github.com/Haiyang-W/CAGroup3D.Comment: Accept by NeurIPS202

    Small Object Detection Based on Two-Stage Calculation Transformer

    Get PDF
    Despite the current small object detection task has achieved significant improvements, it still suffers from some problems. For example, it is a challenge to extract small object features because of little information in the scene of small objects, which may lose the original feature information of small object, resulting in poor detection results. To address this problem, this paper proposes a two-stage calculation Transformer (TCT) based small object detection network. Firstly, a two-stage calculation Transformer is embedded in the backbone feature extraction network for feature enhancement. Based on the traditional Transformer values computation, multiple 1D dilated convolutional layer branches with different feature fusions are utilized to implement global self-attention for the purpose of improving the feature representation and information interaction. Secondly, this paper proposes an effective residual connection module to improve the low-efficiency convolution and activation of the current CSPLayer, which helps to advance the information flow and learn more rich contextual details. Finally, this paper proposes a feature fusion and refinement module for fusing multi-scale features and improving the target feature representation capability. Quantitative and qualitative experiments on PASCAL VOC2007+2012 dataset, COCO2017 dataset and TinyPerson dataset show that the proposed algorithm has better ability of target feature extraction and higher detection accuracy for small target detection, compared with YOLOX

    Advances of deep learning in electrical impedance tomography image reconstruction

    Get PDF
    Electrical impedance tomography (EIT) has been widely used in biomedical research because of its advantages of real-time imaging and nature of being non-invasive and radiation-free. Additionally, it can reconstruct the distribution or changes in electrical properties in the sensing area. Recently, with the significant advancements in the use of deep learning in intelligent medical imaging, EIT image reconstruction based on deep learning has received considerable attention. This study introduces the basic principles of EIT and summarizes the application progress of deep learning in EIT image reconstruction with regards to three aspects: a single network reconstruction, deep learning combined with traditional algorithm reconstruction, and multiple network hybrid reconstruction. In future, optimizing the datasets may be the main challenge in applying deep learning for EIT image reconstruction. Adopting a better network structure, focusing on the joint reconstruction of EIT and traditional algorithms, and using multimodal deep learning-based EIT may be the solution to existing problems. In general, deep learning offers a fresh approach for improving the performance of EIT image reconstruction and could be the foundation for building an intelligent integrated EIT diagnostic system in the future

    Resistin induces multidrug resistance in myeloma by inhibiting cell death and upregulating ABC transporter expression

    Get PDF
    Despite advances in therapy, multiple myeloma remains incurable, with a high frequency of relapse. This suggests the need to identify additional factors that contribute to drug resistance. Our previous studies revealed that bone marrow adipocytes promote resistance to chemotherapy in myeloma through adipocyte-secreted adipokines, but the mechanism underlying this effect and the specific adipokines involved are not well understood. We proposed to determine the role of resistin, an adipokine that is secreted by adipocytes, in chemotherapy resistance in myeloma. We found that resistin abrogated chemotherapy-induced apoptosis in established myeloma cell lines and primary myeloma samples. Resistin inhibited chemotherapy-induced caspase cleavage through the NF-κB and PI3K/Akt pathways. Resistin also increased the expression and drug efflux function of ATP-binding cassette (ABC) transporters in myeloma cells through decreasing the expression of both DNA methyltransferases DNMT1 and DNMT3a and the methylation levels of ABC gene promoters. In vivo studies further demonstrated the protective effect of resistin in chemotherapy-induced apoptosis. Our study thus reveals a new biological function of resistin in the pathogenesis of myeloma, with the implication that targeting resistin could be a potential strategy to prevent or overcome multidrug resistance in myeloma
    corecore