328 research outputs found
Adjectives in Qiang
Qiang is a Tibeto-Burman language spoken by 70,000-80,000 people in Northern Sichuan Province, China, classified as being in the Qiang or Tibetan nationality by the Chinese government. The language is verb final, agglutinative (prefixing and suffixing), and has both head-marking and dependent-marking morphology
The copula and existential verbs in Qiang
This paper discusses the copula and existential verb constructions in Qiang, a Tibeto-Burman language of northern Sichuan, China
A novel machine-vision-based facility for the automatic evaluation of yield-related traits in rice
The evaluation of yield-related traits is an essential step in rice breeding, genetic research and functional genomics research. A new, automatic, and labor-free facility to automatically thresh rice panicles, evaluate rice yield traits, and subsequently pack filled spikelets is presented in this paper. Tests showed that the facility was capable of evaluating yield-related traits with a mean absolute percentage error of less than 5% and an efficiency of 1440 plants per continuous 24 h workday
RGBT Tracking via Progressive Fusion Transformer with Dynamically Guided Learning
Existing Transformer-based RGBT tracking methods either use cross-attention
to fuse the two modalities, or use self-attention and cross-attention to model
both modality-specific and modality-sharing information. However, the
significant appearance gap between modalities limits the feature representation
ability of certain modalities during the fusion process. To address this
problem, we propose a novel Progressive Fusion Transformer called ProFormer,
which progressively integrates single-modality information into the multimodal
representation for robust RGBT tracking. In particular, ProFormer first uses a
self-attention module to collaboratively extract the multimodal representation,
and then uses two cross-attention modules to interact it with the features of
the dual modalities respectively. In this way, the modality-specific
information can well be activated in the multimodal representation. Finally, a
feed-forward network is used to fuse two interacted multimodal representations
for the further enhancement of the final multimodal representation. In
addition, existing learning methods of RGBT trackers either fuse multimodal
features into one for final classification, or exploit the relationship between
unimodal branches and fused branch through a competitive learning strategy.
However, they either ignore the learning of single-modality branches or result
in one branch failing to be well optimized. To solve these problems, we propose
a dynamically guided learning algorithm that adaptively uses well-performing
branches to guide the learning of other branches, for enhancing the
representation ability of each branch. Extensive experiments demonstrate that
our proposed ProFormer sets a new state-of-the-art performance on RGBT210,
RGBT234, LasHeR, and VTUAV datasets.Comment: 13 pages, 9 figure
- …