676 research outputs found
Understanding the Roles of Different Transport Modes in Logistics Market: Content Analysis for an Online Logistics Forum
For the purpose of meeting customer requirements at minimum cost, different transport modes need to be coordinated to make full use of their respective advantages in logistics market. A critical challenge lies in the lack of understanding for the roles of different transport modes in the dynamic logistics market with uncertainties. Online logistics forums provide user-generated content representing real-time market information. In this paper, a content analysis based method is presented to explore the logistics market. Text content in logistics forums is processed by word segmentation and stop word filtering. Then the LDA topic model is derived representing the most probable words for each transport mode. On this basis, the market areas and the cargo types can be investigated for the different modes of transportation. The method is demonstrated using a case study
Predictors of recovery from post-traumatic stress disorder after the dongting lake flood in China: a 13–14 year follow-up study
Influenza nucleoprotein delivered with aluminium salts protects mice from an influenza virus that expresses an altered nucleoprotein sequence
Influenza virus poses a difficult challenge for protective immunity. This virus is adept at altering its surface proteins, the proteins that are the targets of neutralizing antibody. Consequently, each year a new vaccine must be developed to combat the current recirculating strains. A universal influenza vaccine that primes specific memory cells that recognise conserved parts of the virus could prove to be effective against both annual influenza variants and newly emergent potentially pandemic strains. Such a vaccine will have to contain a safe and effective adjuvant that can be used in individuals of all ages. We examine protection from viral challenge in mice vaccinated with the nucleoprotein from the PR8 strain of influenza A, a protein that is highly conserved across viral subtypes. Vaccination with nucleoprotein delivered with a universally used and safe adjuvant, composed of insoluble aluminium salts, provides protection against viruses that either express the same or an altered version of nucleoprotein. This protection correlated with the presence of nucleoprotein specific CD8 T cells in the lungs of infected animals at early time points after infection. In contrast, immunization with NP delivered with alum and the detoxified LPS adjuvant, monophosphoryl lipid A, provided some protection to the homologous viral strain but no protection against infection by influenza expressing a variant nucleoprotein. Together, these data point towards a vaccine solution for all influenza A subtypes
Reconstruction of compressed spectral imaging based on global structure and spectral correlation
In this paper, a convolution sparse coding method based on global structure
characteristics and spectral correlation is proposed for the reconstruction of
compressive spectral images. The proposed method uses the convolution kernel to
operate the global image, which can better preserve image structure information
in the spatial dimension. To take full exploration of the constraints between
spectra, the coefficients corresponding to the convolution kernel are
constrained by the norm to improve spectral accuracy. And, to solve the problem
that convolutional sparse coding is insensitive to low frequency, the global
total-variation (TV) constraint is added to estimate the low-frequency
components. It not only ensures the effective estimation of the low-frequency
but also transforms the convolutional sparse coding into a de-noising process,
which makes the reconstructing process simpler. Simulations show that compared
with the current mainstream optimization methods (DeSCI and Gap-TV), the
proposed method improves the reconstruction quality by up to 7 dB in PSNR and
10% in SSIM, and has a great improvement in the details of the reconstructed
image
Open-world Story Generation with Structured Knowledge Enhancement: A Comprehensive Survey
Storytelling and narrative are fundamental to human experience, intertwined
with our social and cultural engagement. As such, researchers have long
attempted to create systems that can generate stories automatically. In recent
years, powered by deep learning and massive data resources, automatic story
generation has shown significant advances. However, considerable challenges,
like the need for global coherence in generated stories, still hamper
generative models from reaching the same storytelling ability as human
narrators. To tackle these challenges, many studies seek to inject structured
knowledge into the generation process, which is referred to as structure
knowledge-enhanced story generation. Incorporating external knowledge can
enhance the logical coherence among story events, achieve better knowledge
grounding, and alleviate over-generalization and repetition problems in
stories. This survey provides the latest and comprehensive review of this
research field: (i) we present a systematical taxonomy regarding how existing
methods integrate structured knowledge into story generation; (ii) we summarize
involved story corpora, structured knowledge datasets, and evaluation metrics;
(iii) we give multidimensional insights into the challenges of
knowledge-enhanced story generation and cast light on promising directions for
future study
3D-Aware Neural Body Fitting for Occlusion Robust 3D Human Pose Estimation
Regression-based methods for 3D human pose estimation directly predict the 3D
pose parameters from a 2D image using deep networks. While achieving
state-of-the-art performance on standard benchmarks, their performance degrades
under occlusion. In contrast, optimization-based methods fit a parametric body
model to 2D features in an iterative manner. The localized reconstruction loss
can potentially make them robust to occlusion, but they suffer from the 2D-3D
ambiguity.
Motivated by the recent success of generative models in rigid object pose
estimation, we propose 3D-aware Neural Body Fitting (3DNBF) - an approximate
analysis-by-synthesis approach to 3D human pose estimation with SOTA
performance and occlusion robustness. In particular, we propose a generative
model of deep features based on a volumetric human representation with Gaussian
ellipsoidal kernels emitting 3D pose-dependent feature vectors. The neural
features are trained with contrastive learning to become 3D-aware and hence to
overcome the 2D-3D ambiguity.
Experiments show that 3DNBF outperforms other approaches on both occluded and
standard benchmarks. Code is available at https://github.com/edz-o/3DNBFComment: ICCV 2023, project page: https://3dnbf.github.io
FedConv: Enhancing Convolutional Neural Networks for Handling Data Heterogeneity in Federated Learning
Federated learning (FL) is an emerging paradigm in machine learning, where a
shared model is collaboratively learned using data from multiple devices to
mitigate the risk of data leakage. While recent studies posit that Vision
Transformer (ViT) outperforms Convolutional Neural Networks (CNNs) in
addressing data heterogeneity in FL, the specific architectural components that
underpin this advantage have yet to be elucidated. In this paper, we
systematically investigate the impact of different architectural elements, such
as activation functions and normalization layers, on the performance within
heterogeneous FL. Through rigorous empirical analyses, we are able to offer the
first-of-its-kind general guidance on micro-architecture design principles for
heterogeneous FL.
Intriguingly, our findings indicate that with strategic architectural
modifications, pure CNNs can achieve a level of robustness that either matches
or even exceeds that of ViTs when handling heterogeneous data clients in FL.
Additionally, our approach is compatible with existing FL techniques and
delivers state-of-the-art solutions across a broad spectrum of FL benchmarks.
The code is publicly available at https://github.com/UCSC-VLAA/FedConvComment: 9 pages, 6 figures. Equal contribution by P. Xu and Z. Wan
In Defense of Image Pre-Training for Spatiotemporal Recognition
Image pre-training, the current de-facto paradigm for a wide range of visual
tasks, is generally less favored in the field of video recognition. By
contrast, a common strategy is to directly train with spatiotemporal
convolutional neural networks (CNNs) from scratch. Nonetheless, interestingly,
by taking a closer look at these from-scratch learned CNNs, we note there exist
certain 3D kernels that exhibit much stronger appearance modeling ability than
others, arguably suggesting appearance information is already well disentangled
in learning. Inspired by this observation, we hypothesize that the key to
effectively leveraging image pre-training lies in the decomposition of learning
spatial and temporal features, and revisiting image pre-training as the
appearance prior to initializing 3D kernels. In addition, we propose
Spatial-Temporal Separable (STS) convolution, which explicitly splits the
feature channels into spatial and temporal groups, to further enable a more
thorough decomposition of spatiotemporal features for fine-tuning 3D CNNs. Our
experiments show that simply replacing 3D convolution with STS notably improves
a wide range of 3D CNNs without increasing parameters and computation on both
Kinetics-400 and Something-Something V2. Moreover, this new training pipeline
consistently achieves better results on video recognition with significant
speedup. For instance, we achieve +0.6% top-1 of Slowfast on Kinetics-400 over
the strong 256-epoch 128-GPU baseline while fine-tuning for only 50 epochs with
4 GPUs. The code and models are available at
https://github.com/UCSC-VLAA/Image-Pretraining-for-Video.Comment: Published as a conference paper at ECCV 202
Mamba-R: Vision Mamba ALSO Needs Registers
Similar to Vision Transformers, this paper identifies artifacts also present
within the feature maps of Vision Mamba. These artifacts, corresponding to
high-norm tokens emerging in low-information background areas of images, appear
much more severe in Vision Mamba -- they exist prevalently even with the
tiny-sized model and activate extensively across background regions. To
mitigate this issue, we follow the prior solution of introducing register
tokens into Vision Mamba. To better cope with Mamba blocks' uni-directional
inference paradigm, two key modifications are introduced: 1) evenly inserting
registers throughout the input token sequence, and 2) recycling registers for
final decision predictions. We term this new architecture Mamba-R. Qualitative
observations suggest, compared to vanilla Vision Mamba, Mamba-R's feature maps
appear cleaner and more focused on semantically meaningful regions.
Quantitatively, Mamba-R attains stronger performance and scales better. For
example, on the ImageNet benchmark, our base-size Mamba-R attains 82.9%
accuracy, significantly outperforming Vim-B's 81.8%; furthermore, we provide
the first successful scaling to the large model size (i.e., with 341M
parameters), attaining a competitive accuracy of 83.2% (84.5% if finetuned with
384x384 inputs). Additional validation on the downstream semantic segmentation
task also supports Mamba-R's efficacy
- …
