162 research outputs found
Geometry-Aware Face Completion and Editing
Face completion is a challenging generation task because it requires
generating visually pleasing new pixels that are semantically consistent with
the unmasked face region. This paper proposes a geometry-aware Face Completion
and Editing NETwork (FCENet) by systematically studying facial geometry from
the unmasked region. Firstly, a facial geometry estimator is learned to
estimate facial landmark heatmaps and parsing maps from the unmasked face
image. Then, an encoder-decoder structure generator serves to complete a face
image and disentangle its mask areas conditioned on both the masked face image
and the estimated facial geometry images. Besides, since low-rank property
exists in manually labeled masks, a low-rank regularization term is imposed on
the disentangled masks, enforcing our completion network to manage occlusion
area with various shape and size. Furthermore, our network can generate diverse
results from the same masked input by modifying estimated facial geometry,
which provides a flexible mean to edit the completed face appearance. Extensive
experimental results qualitatively and quantitatively demonstrate that our
network is able to generate visually pleasing face completion results and edit
face attributes as well
Pruning Large Language Models via Accuracy Predictor
Large language models(LLMs) containing tens of billions of parameters (or
even more) have demonstrated impressive capabilities in various NLP tasks.
However, substantial model size poses challenges to training, inference, and
deployment so that it is necessary to compress the model. At present, most
model compression for LLMs requires manual design of pruning features, which
has problems such as complex optimization pipeline and difficulty in retaining
the capabilities of certain parts of the model.Therefore, we propose a novel
pruning approach: firstly, a training set of a certain number of
architecture-accuracy pairs is established, and then a non-neural model is
trained as an accuracy predictor. Using the accuracy predictor to further
optimize the search space and search, the optimal model can be automatically
selected. Experiments show that our proposed approach is effective and
efficient. Compared with the baseline, the perplexity(PPL) on Wikitext2 and PTB
dropped by 9.48% and 5,76% respectively, and the average accuracy of MMLU
increased by 6.28%.Comment: 6 pages, 4 figs, submitted to IEEE ICASSP 202
Prototype as Query for Few Shot Semantic Segmentation
Few-shot Semantic Segmentation (FSS) was proposed to segment unseen classes
in a query image, referring to only a few annotated examples named support
images. One of the characteristics of FSS is spatial inconsistency between
query and support targets, e.g., texture or appearance. This greatly challenges
the generalization ability of methods for FSS, which requires to effectively
exploit the dependency of the query image and the support examples. Most
existing methods abstracted support features into prototype vectors and
implemented the interaction with query features using cosine similarity or
feature concatenation. However, this simple interaction may not capture spatial
details in query features. To alleviate this limitation, a few methods utilized
all pixel-wise support information via computing the pixel-wise correlations
between paired query and support features implemented with the attention
mechanism of Transformer. These approaches suffer from heavy computation on the
dot-product attention between all pixels of support and query features. In this
paper, we propose a simple yet effective framework built upon Transformer
termed as ProtoFormer to fully capture spatial details in query features. It
views the abstracted prototype of the target class in support features as Query
and the query features as Key and Value embeddings, which are input to the
Transformer decoder. In this way, the spatial details can be better captured
and the semantic features of target class in the query image can be focused.
The output of the Transformer-based module can be viewed as semantic-aware
dynamic kernels to filter out the segmentation mask from the enriched query
features. Extensive experiments on PASCAL- and COCO- show that
our ProtoFormer significantly advances the state-of-the-art methods.Comment: under revie
Vaccine delivery alerts innate immune systems for more immunogenic vaccination
Vaccine delivery technologies are mainly designed to minimally invasively deliver vaccines to target tissues with little or no adjuvant effects. This study presents a prototype laser-based powder delivery (LPD) with inherent adjuvant effects for more immunogenic vaccination without incorporation of external adjuvants. LPD takes advantage of aesthetic ablative fractional laser to generate skin microchannels to support high-efficient vaccine delivery and at the same time creates photothermal stress in microchannel-surrounding tissues to boost vaccination. LPD could significantly enhance pandemic influenza 2009 H1N1 vaccine immunogenicity and protective efficacy as compared with needle-based intradermal delivery in murine models. The ablative fractional laser was found to induce host DNA release, activate NLR family pyrin domain containing 3 inflammasome, and stimulate IL-1β release despite their dispensability for laser adjuvant effects. Instead, the ablative fractional laser activated MyD88 to mediate its adjuvant effects by potentiation of antigen uptake, maturation, and migration of dendritic cells. LPD also induced minimal local or systemic adverse reactions due to the microfractional and sustained vaccine delivery. Our data support the development of self-adjuvanted vaccine delivery technologies by intentional induction of well-controlled tissue stress to alert innate immune systems for more immunogenic vaccination
- …