51 research outputs found
Effects of Flow Compressibility on Two-Phase Mixing in Supersonic Droplet-Laden Flows
This research addresses a numerical analysis on the effects of flow compressibility on the characteristics of droplet dispersion, evaporation, and mixing of fuel and air according to the simulation of the spatially developing supersonic shear flows laden with evaporating n-decane droplets. A sixth-order hybrid WENO numerical scheme is employed for capturing the unsteady wave structures. The influence of inflow convective Mach number (Mc), representing the high-speed flow compressibility, on the twophase mixing is analyzed, in which Mc is specified from 0.4 to 1.0. It is found that the shearing vortex is compressed spatially as Mc increases, associated with the alternate distributions of compression and expansion regimes in the flow field. The flow compressibility changes not only the vortex structures but also the aerothermal parameters of the shear flows, and further influences the dispersion and evaporation of droplets. The two-phase mixing efficiency is observed to decrease as Mc increases
Multi-dimensional Fusion and Consistency for Semi-supervised Medical Image Segmentation
In this paper, we introduce a novel semi-supervised learning framework
tailored for medical image segmentation. Central to our approach is the
innovative Multi-scale Text-aware ViT-CNN Fusion scheme. This scheme adeptly
combines the strengths of both ViTs and CNNs, capitalizing on the unique
advantages of both architectures as well as the complementary information in
vision-language modalities. Further enriching our framework, we propose the
Multi-Axis Consistency framework for generating robust pseudo labels, thereby
enhancing the semi-supervised learning process. Our extensive experiments on
several widely-used datasets unequivocally demonstrate the efficacy of our
approach
Human Pose Driven Object Effects Recommendation
In this paper, we research the new topic of object effects recommendation in
micro-video platforms, which is a challenging but important task for many
practical applications such as advertisement insertion. To avoid the problem of
introducing background bias caused by directly learning video content from
image frames, we propose to utilize the meaningful body language hidden in 3D
human pose for recommendation. To this end, in this work, a novel human pose
driven object effects recommendation network termed PoseRec is introduced.
PoseRec leverages the advantages of 3D human pose detection and learns
information from multi-frame 3D human pose for video-item registration,
resulting in high quality object effects recommendation performance. Moreover,
to solve the inherent ambiguity and sparsity issues that exist in object
effects recommendation, we further propose a novel item-aware implicit
prototype learning module and a novel pose-aware transductive hard-negative
mining module to better learn pose-item relationships. What's more, to
benchmark methods for the new research topic, we build a new dataset for object
effects recommendation named Pose-OBE. Extensive experiments on Pose-OBE
demonstrate that our method can achieve superior performance than strong
baselines
CaSR Induces Osteoclast Differentiation and Promotes Bone Metastasis in Lung Adenocarcinoma
Objective: Explore the mechanism of CaSR's involvement in bone metastasis in lung adenocarcinoma.
Methods: Immunohistochemistry (IHC) was used to detect the expression of calcium-sensing receptor (CaSR) in 120 cases of lung adenocarcinoma with bone metastasis. Stably transfected cell lines with CaSR overexpression and knockdown based on A549 cells were constructed. The expression of CaSR was verified by western blot and qPCR. The proliferation and migration abilities of A549 cells were tested using cholecystokinin-8 (CCK-8) and Transwell assays, respectively. Western blotting was used to detect the expression of matrix metalloproteinases MMP2, MMP9, CaSR, and NF-κB. The supernatant from each cell culture group was collected as a conditional co-culture solution to study the induction of osteoclast precursor cells and osteoblasts. Western blot and qPCR were used to validate the expression of bone matrix degradation-related enzymes cathepsin K and hormone calcitonin receptor (CTR) and osteoblast-induced osteoclast maturation and differentiation enzyme receptor activator of nuclear factor-κB ligand (RANKL), macrophage colony-stimulating factor (M-CSF), osteoprotegerin (OPG), and PTHrP. Immunofluorescent staining was used to detect F-actin ring formation and osteocalcin expression. Western blot results for NF-κB expression identified a regulatory relationship between NF-κB and CaSR.
Results: CaSR expression in lung cancer tissues was significantly higher than that in adjacent and normal lung tissues. The expression of CaSR in lung cancer tissues with bone metastasis was higher than that in non-metastatic lung cancer tissues. The proliferation and migration ability of A549 cells increased significantly with overexpressed CaSR. The co-culture solution directly induced osteoclast precursor cells and the expression of bone matrix degradation-related enzymes significantly increased. Osteoblasts were significantly inhibited and osteoblast-induced osteoclast maturation and differentiation enzymes were significantly downregulated. It was found that the expression of NF-κB and PTHrP increased when CaSR was overexpressed. Osteoclast differentiation factor expression was also significantly increased, which directly induces osteoclast differentiation and maturation. These results were reversed when CaSR was knocked down.
Conclusions: CaSR can positively regulate NF-κB and PTHrP expression in A549 cells with a high metastatic potential, thereby promoting osteoclast differentiation and maturation, and facilitating the occurrence and development of bone metastasis in lung adenocarcinoma
AS-FIBA: Adaptive Selective Frequency-Injection for Backdoor Attack on Deep Face Restoration
Deep learning-based face restoration models, increasingly prevalent in smart
devices, have become targets for sophisticated backdoor attacks. These attacks,
through subtle trigger injection into input face images, can lead to unexpected
restoration outcomes. Unlike conventional methods focused on classification
tasks, our approach introduces a unique degradation objective tailored for
attacking restoration models. Moreover, we propose the Adaptive Selective
Frequency Injection Backdoor Attack (AS-FIBA) framework, employing a neural
network for input-specific trigger generation in the frequency domain,
seamlessly blending triggers with benign images. This results in imperceptible
yet effective attacks, guiding restoration predictions towards subtly degraded
outputs rather than conspicuous targets. Extensive experiments demonstrate the
efficacy of the degradation objective on state-of-the-art face restoration
models. Additionally, it is notable that AS-FIBA can insert effective backdoors
that are more imperceptible than existing backdoor attack methods, including
WaNet, ISSBA, and FIBA
D-IF: Uncertainty-aware Human Digitization via Implicit Distribution Field
Realistic virtual humans play a crucial role in numerous industries, such as
metaverse, intelligent healthcare, and self-driving simulation. But creating
them on a large scale with high levels of realism remains a challenge. The
utilization of deep implicit function sparks a new era of image-based 3D
clothed human reconstruction, enabling pixel-aligned shape recovery with fine
details. Subsequently, the vast majority of works locate the surface by
regressing the deterministic implicit value for each point. However, should all
points be treated equally regardless of their proximity to the surface? In this
paper, we propose replacing the implicit value with an adaptive uncertainty
distribution, to differentiate between points based on their distance to the
surface. This simple ``value to distribution'' transition yields significant
improvements on nearly all the baselines. Furthermore, qualitative results
demonstrate that the models trained using our uncertainty distribution loss,
can capture more intricate wrinkles, and realistic limbs. Code and models are
available for research purposes at https://github.com/psyai-net/D-IF_release
Reconstruction-Aware Prior Distillation for Semi-supervised Point Cloud Completion
Point clouds scanned by real-world sensors are always incomplete, irregular,
and noisy, making the point cloud completion task become increasingly more
important. Though many point cloud completion methods have been proposed, most
of them require a large number of paired complete-incomplete point clouds for
training, which is labor exhausted. In contrast, this paper proposes a novel
Reconstruction-Aware Prior Distillation semi-supervised point cloud completion
method named RaPD, which takes advantage of a two-stage training scheme to
reduce the dependence on a large-scale paired dataset. In training stage 1, the
so-called deep semantic prior is learned from both unpaired complete and
unpaired incomplete point clouds using a reconstruction-aware pretraining
process. While in training stage 2, we introduce a semi-supervised prior
distillation process, where an encoder-decoder-based completion network is
trained by distilling the prior into the network utilizing only a small number
of paired training samples. A self-supervised completion module is further
introduced, excavating the value of a large number of unpaired incomplete point
clouds, leading to an increase in the network's performance. Extensive
experiments on several widely used datasets demonstrate that RaPD, the first
semi-supervised point cloud completion method, achieves superior performance to
previous methods on both homologous and heterologous scenarios
Object Level Depth Reconstruction for Category Level 6D Object Pose Estimation From Monocular RGB Image
Recently, RGBD-based category-level 6D object pose estimation has achieved
promising improvement in performance, however, the requirement of depth
information prohibits broader applications. In order to relieve this problem,
this paper proposes a novel approach named Object Level Depth reconstruction
Network (OLD-Net) taking only RGB images as input for category-level 6D object
pose estimation. We propose to directly predict object-level depth from a
monocular RGB image by deforming the category-level shape prior into
object-level depth and the canonical NOCS representation. Two novel modules
named Normalized Global Position Hints (NGPH) and Shape-aware Decoupled Depth
Reconstruction (SDDR) module are introduced to learn high fidelity object-level
depth and delicate shape representations. At last, the 6D object pose is solved
by aligning the predicted canonical representation with the back-projected
object-level depth. Extensive experiments on the challenging CAMERA25 and
REAL275 datasets indicate that our model, though simple, achieves
state-of-the-art performance.Comment: 19 pages, 7 figures, 4 table
EmoTalk: Speech-Driven Emotional Disentanglement for 3D Face Animation
Speech-driven 3D face animation aims to generate realistic facial expressions
that match the speech content and emotion. However, existing methods often
neglect emotional facial expressions or fail to disentangle them from speech
content. To address this issue, this paper proposes an end-to-end neural
network to disentangle different emotions in speech so as to generate rich 3D
facial expressions. Specifically, we introduce the emotion disentangling
encoder (EDE) to disentangle the emotion and content in the speech by
cross-reconstructed speech signals with different emotion labels. Then an
emotion-guided feature fusion decoder is employed to generate a 3D talking face
with enhanced emotion. The decoder is driven by the disentangled identity,
emotional, and content embeddings so as to generate controllable personal and
emotional styles. Finally, considering the scarcity of the 3D emotional talking
face data, we resort to the supervision of facial blendshapes, which enables
the reconstruction of plausible 3D faces from 2D emotional data, and contribute
a large-scale 3D emotional talking face dataset (3D-ETF) to train the network.
Our experiments and user studies demonstrate that our approach outperforms
state-of-the-art methods and exhibits more diverse facial movements. We
recommend watching the supplementary video:
https://ziqiaopeng.github.io/emotalkComment: Accepted by ICCV 202
- …