701 research outputs found

    Efficient Adaptation of Large Vision Transformer via Adapter Re-Composing

    Full text link
    The advent of high-capacity pre-trained models has revolutionized problem-solving in computer vision, shifting the focus from training task-specific models to adapting pre-trained models. Consequently, effectively adapting large pre-trained models to downstream tasks in an efficient manner has become a prominent research area. Existing solutions primarily concentrate on designing lightweight adapters and their interaction with pre-trained models, with the goal of minimizing the number of parameters requiring updates. In this study, we propose a novel Adapter Re-Composing (ARC) strategy that addresses efficient pre-trained model adaptation from a fresh perspective. Our approach considers the reusability of adaptation parameters and introduces a parameter-sharing scheme. Specifically, we leverage symmetric down-/up-projections to construct bottleneck operations, which are shared across layers. By learning low-dimensional re-scaling coefficients, we can effectively re-compose layer-adaptive adapters. This parameter-sharing strategy in adapter design allows us to significantly reduce the number of new parameters while maintaining satisfactory performance, thereby offering a promising approach to compress the adaptation cost. We conduct experiments on 24 downstream image classification tasks using various Vision Transformer variants to evaluate our method. The results demonstrate that our approach achieves compelling transfer learning performance with a reduced parameter count. Our code is available at \href{https://github.com/DavidYanAnDe/ARC}{https://github.com/DavidYanAnDe/ARC}.Comment: Paper is accepted to NeurIPS 202

    Static stiffness modeling and sensitivity analysis for geared system used for rotary feeding

    Get PDF
    The positioning accuracy of rotary feed system under load greatly depends on the static stiffness of mechanical transmission system. This paper proposes a unified static stiffness model of rotary feed system with geared transmission system. Taking the torsional stiffness of transmission shaft and mesh stiffness of gear pairs into account, the motion equations of the whole transmission system are presented. Based on the static equilibrium, a unified expression for the relationship between torsional angles of two adjacent elements is derived. Then a unified static stiffness model is presented. Furthermore, analytical expressions for sensitivity analysis of the static stiffness on the individual element’s stiffness and design parameters are derived. The presented model is verified by a traditional model, and a good agreement is obtained. The influence of phase angle of meshing gear pairs on the resultant static stiffness is investigated. An example transmission system is employed to perform the sensitivity analysis and the results are analyzed. The proposed model provides an essential tool for the design of rotary feed system satisfying requirement of static stiffness

    Improved Acid Resistance of a Metal-Organic Cage Enables Cargo Release and Exchange between Hosts.

    Get PDF
    The use of di(2-pyridyl)ketone in subcomponent self-assembly is introduced. When combined with a flexible triamine and zinc bis(trifluoromethanesulfonyl)imide, this ketone formed a new Zn4 L4 tetrahedron 1 bearing twelve uncoordinated pyridyl units around its metal-ion vertices. The acid stability of 1 was found to be greater than that of the analogous tetrahedron 2 built from 2-formylpyridine. Intriguingly, the peripheral presence of additional pyridine rings in 1 resulted in distinct guest binding behavior from that of 2, affecting guest scope as well as binding affinities. The different stabilities and guest affinities of capsules 1 and 2 enabled the design of systems whereby different cargoes could be moved between cages using acid and base as chemical stimuli.European Research Council (695009), UK Engineering and Physical Sciences Research Council (EPSRC EP/P027067/1

    Post-Training Quantization for Object Detection

    Full text link
    Efficient inference for object detection networks is a major challenge on edge devices. Post-Training Quantization (PTQ), which transforms a full-precision model into low bit-width directly, is an effective and convenient approach to reduce model inference complexity. But it suffers severe accuracy drop when applied to complex tasks such as object detection. PTQ optimizes the quantization parameters by different metrics to minimize the perturbation of quantization. The p-norm distance of feature maps before and after quantization, Lp, is widely used as the metric to evaluate perturbation. For the specialty of object detection network, we observe that the parameter p in Lp metric will significantly influence its quantization performance. We indicate that using a fixed hyper-parameter p does not achieve optimal quantization performance. To mitigate this problem, we propose a framework, DetPTQ, to assign different p values for quantizing different layers using an Object Detection Output Loss (ODOL), which represents the task loss of object detection. DetPTQ employs the ODOL-based adaptive Lp metric to select the optimal quantization parameters. Experiments show that our DetPTQ outperforms the state-of-the-art PTQ methods by a significant margin on both 2D and 3D object detectors. For example, we achieve 31.1/31.7(quantization/full-precision) mAP on RetinaNet-ResNet18 with 4-bit weight and 4-bit activation

    PD-Quant: Post-Training Quantization based on Prediction Difference Metric

    Full text link
    Post-training quantization (PTQ) is a neural network compression technique that converts a full-precision model into a quantized model using lower-precision data types. Although it can help reduce the size and computational cost of deep neural networks, it can also introduce quantization noise and reduce prediction accuracy, especially in extremely low-bit settings. How to determine the appropriate quantization parameters (e.g., scaling factors and rounding of weights) is the main problem facing now. Existing methods attempt to determine these parameters by minimize the distance between features before and after quantization, but such an approach only considers local information and may not result in the most optimal quantization parameters. We analyze this issue and ropose PD-Quant, a method that addresses this limitation by considering global information. It determines the quantization parameters by using the information of differences between network prediction before and after quantization. In addition, PD-Quant can alleviate the overfitting problem in PTQ caused by the small number of calibration sets by adjusting the distribution of activations. Experiments show that PD-Quant leads to better quantization parameters and improves the prediction accuracy of quantized models, especially in low-bit settings. For example, PD-Quant pushes the accuracy of ResNet-18 up to 53.14% and RegNetX-600MF up to 40.67% in weight 2-bit activation 2-bit. The code is released at https://github.com/hustvl/PD-Quant
    corecore