1,641 research outputs found
TUNet: A Block-online Bandwidth Extension Model based on Transformers and Self-supervised Pretraining
We introduce a block-online variant of the temporal feature-wise linear
modulation (TFiLM) model to achieve bandwidth extension. The proposed
architecture simplifies the UNet backbone of the TFiLM to reduce inference time
and employs an efficient transformer at the bottleneck to alleviate performance
degradation. We also utilize self-supervised pretraining and data augmentation
to enhance the quality of bandwidth extended signals and reduce the sensitivity
with respect to downsampling methods. Experiment results on the VCTK dataset
show that the proposed method outperforms several recent baselines in both
intrusive and non-intrusive metrics. Pretraining and filter augmentation also
help stabilize and enhance the overall performance.Comment: Published as a conference paper at ICASSP 2022, 5 pages, 4 figures, 3
table
Recommended from our members
Justification Patterns for OWL DL Ontologies
For debugging OWL-DL ontologies, natural language explanations of inconsistencies and undesirable entailments are of great help. From such explanations, ontology developers can learn why an ontology gives rise to specific entailments. Unfortunately, commonly used tableaux-based reasoning services do not provide a basis for such explanations, since they rely on a refutation proof strategy and normalising transformations that are difficult for human ontology editors to understand. For this reason, we investigate the use of automatically generated justifications for entailments (i.e., minimal sets of axioms from the ontology that cause entailments to hold). We show that such justifications fall into a manageable number of patterns, which can be used as a basis for generating natural language explanations by associating each justification pattern with a rhetorical pattern in natural language
Channel and spatial attention mechanism for fashion image captioning
Image captioning aims to automatically generate one or more description sentences for a given input image. Most of the existing captioning methods use encoder-decoder model which mainly focus on recognizing and capturing the relationship between objects appearing in the input image. However, when generating captions for fashion images, it is important to not only describe the items and their relationships, but also mention attribute features of clothes (shape, texture, style, fabric, and more). In this study, one novel model is proposed for fashion image captioning task which can capture not only the items and their relationship, but also their attribute features. Two different attention mechanisms (spatial-attention and channel-wise attention) is incorporated to the traditional encoder-decoder model, which dynamically interprets the caption sentence in multi-layer feature map in addition to the depth dimension of the feature map. We evaluate our proposed architecture on Fashion-Gen using three different metrics (CIDEr, ROUGE-L, and BLEU-1), and achieve the scores of 89.7, 50.6 and 45.6, respectively. Based on experiments, our proposed method shows significant performance improvement for the task of fashion-image captioning, and outperforms other state-of-the-art image captioning methods
Conditional Support Alignment for Domain Adaptation with Label Shift
Unsupervised domain adaptation (UDA) refers to a domain adaptation framework
in which a learning model is trained based on the labeled samples on the source
domain and unlabelled ones in the target domain. The dominant existing methods
in the field that rely on the classical covariate shift assumption to learn
domain-invariant feature representation have yielded suboptimal performance
under the label distribution shift between source and target domains. In this
paper, we propose a novel conditional adversarial support alignment (CASA)
whose aim is to minimize the conditional symmetric support divergence between
the source's and target domain's feature representation distributions, aiming
at a more helpful representation for the classification task. We also introduce
a novel theoretical target risk bound, which justifies the merits of aligning
the supports of conditional feature distributions compared to the existing
marginal support alignment approach in the UDA settings. We then provide a
complete training process for learning in which the objective optimization
functions are precisely based on the proposed target risk bound. Our empirical
results demonstrate that CASA outperforms other state-of-the-art methods on
different UDA benchmark tasks under label shift conditions
Irreducible representations of Upq[gl(2/2)]
The two-parametric quantum superalgebra and its
representations are considered. All finite-dimensional irreducible
representations of this quantum superalgebra can be constructed and classified
into typical and nontypical ones according to a proposition proved in the
present paper. This proposition is a nontrivial deformation from the one for
the classical superalgebra gl(2/2), unlike the case of one-parametric
deformations.Comment: Latex, 8 pages. A reference added in v.
Beyond Traditional Approaches: Multi-Task Network for Breast Ultrasound Diagnosis
Breast Ultrasound plays a vital role in cancer diagnosis as a non-invasive
approach with cost-effective. In recent years, with the development of deep
learning, many CNN-based approaches have been widely researched in both tumor
localization and cancer classification tasks. Even though previous single
models achieved great performance in both tasks, these methods have some
limitations in inference time, GPU requirement, and separate fine-tuning for
each model. In this study, we aim to redesign and build end-to-end multi-task
architecture to conduct both segmentation and classification. With our proposed
approach, we achieved outstanding performance and time efficiency, with 79.8%
and 86.4% in DeepLabV3+ architecture in the segmentation task.Comment: 7 pages, 3 figure
- …