162 research outputs found

    Geometry-Aware Face Completion and Editing

    Full text link
    Face completion is a challenging generation task because it requires generating visually pleasing new pixels that are semantically consistent with the unmasked face region. This paper proposes a geometry-aware Face Completion and Editing NETwork (FCENet) by systematically studying facial geometry from the unmasked region. Firstly, a facial geometry estimator is learned to estimate facial landmark heatmaps and parsing maps from the unmasked face image. Then, an encoder-decoder structure generator serves to complete a face image and disentangle its mask areas conditioned on both the masked face image and the estimated facial geometry images. Besides, since low-rank property exists in manually labeled masks, a low-rank regularization term is imposed on the disentangled masks, enforcing our completion network to manage occlusion area with various shape and size. Furthermore, our network can generate diverse results from the same masked input by modifying estimated facial geometry, which provides a flexible mean to edit the completed face appearance. Extensive experimental results qualitatively and quantitatively demonstrate that our network is able to generate visually pleasing face completion results and edit face attributes as well

    Pruning Large Language Models via Accuracy Predictor

    Full text link
    Large language models(LLMs) containing tens of billions of parameters (or even more) have demonstrated impressive capabilities in various NLP tasks. However, substantial model size poses challenges to training, inference, and deployment so that it is necessary to compress the model. At present, most model compression for LLMs requires manual design of pruning features, which has problems such as complex optimization pipeline and difficulty in retaining the capabilities of certain parts of the model.Therefore, we propose a novel pruning approach: firstly, a training set of a certain number of architecture-accuracy pairs is established, and then a non-neural model is trained as an accuracy predictor. Using the accuracy predictor to further optimize the search space and search, the optimal model can be automatically selected. Experiments show that our proposed approach is effective and efficient. Compared with the baseline, the perplexity(PPL) on Wikitext2 and PTB dropped by 9.48% and 5,76% respectively, and the average accuracy of MMLU increased by 6.28%.Comment: 6 pages, 4 figs, submitted to IEEE ICASSP 202

    Prototype as Query for Few Shot Semantic Segmentation

    Full text link
    Few-shot Semantic Segmentation (FSS) was proposed to segment unseen classes in a query image, referring to only a few annotated examples named support images. One of the characteristics of FSS is spatial inconsistency between query and support targets, e.g., texture or appearance. This greatly challenges the generalization ability of methods for FSS, which requires to effectively exploit the dependency of the query image and the support examples. Most existing methods abstracted support features into prototype vectors and implemented the interaction with query features using cosine similarity or feature concatenation. However, this simple interaction may not capture spatial details in query features. To alleviate this limitation, a few methods utilized all pixel-wise support information via computing the pixel-wise correlations between paired query and support features implemented with the attention mechanism of Transformer. These approaches suffer from heavy computation on the dot-product attention between all pixels of support and query features. In this paper, we propose a simple yet effective framework built upon Transformer termed as ProtoFormer to fully capture spatial details in query features. It views the abstracted prototype of the target class in support features as Query and the query features as Key and Value embeddings, which are input to the Transformer decoder. In this way, the spatial details can be better captured and the semantic features of target class in the query image can be focused. The output of the Transformer-based module can be viewed as semantic-aware dynamic kernels to filter out the segmentation mask from the enriched query features. Extensive experiments on PASCAL-5i5^{i} and COCO-20i20^{i} show that our ProtoFormer significantly advances the state-of-the-art methods.Comment: under revie

    Vaccine delivery alerts innate immune systems for more immunogenic vaccination

    Get PDF
    Vaccine delivery technologies are mainly designed to minimally invasively deliver vaccines to target tissues with little or no adjuvant effects. This study presents a prototype laser-based powder delivery (LPD) with inherent adjuvant effects for more immunogenic vaccination without incorporation of external adjuvants. LPD takes advantage of aesthetic ablative fractional laser to generate skin microchannels to support high-efficient vaccine delivery and at the same time creates photothermal stress in microchannel-surrounding tissues to boost vaccination. LPD could significantly enhance pandemic influenza 2009 H1N1 vaccine immunogenicity and protective efficacy as compared with needle-based intradermal delivery in murine models. The ablative fractional laser was found to induce host DNA release, activate NLR family pyrin domain containing 3 inflammasome, and stimulate IL-1β release despite their dispensability for laser adjuvant effects. Instead, the ablative fractional laser activated MyD88 to mediate its adjuvant effects by potentiation of antigen uptake, maturation, and migration of dendritic cells. LPD also induced minimal local or systemic adverse reactions due to the microfractional and sustained vaccine delivery. Our data support the development of self-adjuvanted vaccine delivery technologies by intentional induction of well-controlled tissue stress to alert innate immune systems for more immunogenic vaccination
    • …
    corecore