51 research outputs found

    Effects of Flow Compressibility on Two-Phase Mixing in Supersonic Droplet-Laden Flows

    Get PDF
    This research addresses a numerical analysis on the effects of flow compressibility on the characteristics of droplet dispersion, evaporation, and mixing of fuel and air according to the simulation of the spatially developing supersonic shear flows laden with evaporating n-decane droplets. A sixth-order hybrid WENO numerical scheme is employed for capturing the unsteady wave structures. The influence of inflow convective Mach number (Mc), representing the high-speed flow compressibility, on the twophase mixing is analyzed, in which Mc is specified from 0.4 to 1.0. It is found that the shearing vortex is compressed spatially as Mc increases, associated with the alternate distributions of compression and expansion regimes in the flow field. The flow compressibility changes not only the vortex structures but also the aerothermal parameters of the shear flows, and further influences the dispersion and evaporation of droplets. The two-phase mixing efficiency is observed to decrease as Mc increases

    Multi-dimensional Fusion and Consistency for Semi-supervised Medical Image Segmentation

    Full text link
    In this paper, we introduce a novel semi-supervised learning framework tailored for medical image segmentation. Central to our approach is the innovative Multi-scale Text-aware ViT-CNN Fusion scheme. This scheme adeptly combines the strengths of both ViTs and CNNs, capitalizing on the unique advantages of both architectures as well as the complementary information in vision-language modalities. Further enriching our framework, we propose the Multi-Axis Consistency framework for generating robust pseudo labels, thereby enhancing the semi-supervised learning process. Our extensive experiments on several widely-used datasets unequivocally demonstrate the efficacy of our approach

    Human Pose Driven Object Effects Recommendation

    Full text link
    In this paper, we research the new topic of object effects recommendation in micro-video platforms, which is a challenging but important task for many practical applications such as advertisement insertion. To avoid the problem of introducing background bias caused by directly learning video content from image frames, we propose to utilize the meaningful body language hidden in 3D human pose for recommendation. To this end, in this work, a novel human pose driven object effects recommendation network termed PoseRec is introduced. PoseRec leverages the advantages of 3D human pose detection and learns information from multi-frame 3D human pose for video-item registration, resulting in high quality object effects recommendation performance. Moreover, to solve the inherent ambiguity and sparsity issues that exist in object effects recommendation, we further propose a novel item-aware implicit prototype learning module and a novel pose-aware transductive hard-negative mining module to better learn pose-item relationships. What's more, to benchmark methods for the new research topic, we build a new dataset for object effects recommendation named Pose-OBE. Extensive experiments on Pose-OBE demonstrate that our method can achieve superior performance than strong baselines

    CaSR Induces Osteoclast Differentiation and Promotes Bone Metastasis in Lung Adenocarcinoma

    Get PDF
    Objective: Explore the mechanism of CaSR's involvement in bone metastasis in lung adenocarcinoma. Methods: Immunohistochemistry (IHC) was used to detect the expression of calcium-sensing receptor (CaSR) in 120 cases of lung adenocarcinoma with bone metastasis. Stably transfected cell lines with CaSR overexpression and knockdown based on A549 cells were constructed. The expression of CaSR was verified by western blot and qPCR. The proliferation and migration abilities of A549 cells were tested using cholecystokinin-8 (CCK-8) and Transwell assays, respectively. Western blotting was used to detect the expression of matrix metalloproteinases MMP2, MMP9, CaSR, and NF-κB. The supernatant from each cell culture group was collected as a conditional co-culture solution to study the induction of osteoclast precursor cells and osteoblasts. Western blot and qPCR were used to validate the expression of bone matrix degradation-related enzymes cathepsin K and hormone calcitonin receptor (CTR) and osteoblast-induced osteoclast maturation and differentiation enzyme receptor activator of nuclear factor-κB ligand (RANKL), macrophage colony-stimulating factor (M-CSF), osteoprotegerin (OPG), and PTHrP. Immunofluorescent staining was used to detect F-actin ring formation and osteocalcin expression. Western blot results for NF-κB expression identified a regulatory relationship between NF-κB and CaSR. Results: CaSR expression in lung cancer tissues was significantly higher than that in adjacent and normal lung tissues. The expression of CaSR in lung cancer tissues with bone metastasis was higher than that in non-metastatic lung cancer tissues. The proliferation and migration ability of A549 cells increased significantly with overexpressed CaSR. The co-culture solution directly induced osteoclast precursor cells and the expression of bone matrix degradation-related enzymes significantly increased. Osteoblasts were significantly inhibited and osteoblast-induced osteoclast maturation and differentiation enzymes were significantly downregulated. It was found that the expression of NF-κB and PTHrP increased when CaSR was overexpressed. Osteoclast differentiation factor expression was also significantly increased, which directly induces osteoclast differentiation and maturation. These results were reversed when CaSR was knocked down. Conclusions: CaSR can positively regulate NF-κB and PTHrP expression in A549 cells with a high metastatic potential, thereby promoting osteoclast differentiation and maturation, and facilitating the occurrence and development of bone metastasis in lung adenocarcinoma

    AS-FIBA: Adaptive Selective Frequency-Injection for Backdoor Attack on Deep Face Restoration

    Full text link
    Deep learning-based face restoration models, increasingly prevalent in smart devices, have become targets for sophisticated backdoor attacks. These attacks, through subtle trigger injection into input face images, can lead to unexpected restoration outcomes. Unlike conventional methods focused on classification tasks, our approach introduces a unique degradation objective tailored for attacking restoration models. Moreover, we propose the Adaptive Selective Frequency Injection Backdoor Attack (AS-FIBA) framework, employing a neural network for input-specific trigger generation in the frequency domain, seamlessly blending triggers with benign images. This results in imperceptible yet effective attacks, guiding restoration predictions towards subtly degraded outputs rather than conspicuous targets. Extensive experiments demonstrate the efficacy of the degradation objective on state-of-the-art face restoration models. Additionally, it is notable that AS-FIBA can insert effective backdoors that are more imperceptible than existing backdoor attack methods, including WaNet, ISSBA, and FIBA

    D-IF: Uncertainty-aware Human Digitization via Implicit Distribution Field

    Full text link
    Realistic virtual humans play a crucial role in numerous industries, such as metaverse, intelligent healthcare, and self-driving simulation. But creating them on a large scale with high levels of realism remains a challenge. The utilization of deep implicit function sparks a new era of image-based 3D clothed human reconstruction, enabling pixel-aligned shape recovery with fine details. Subsequently, the vast majority of works locate the surface by regressing the deterministic implicit value for each point. However, should all points be treated equally regardless of their proximity to the surface? In this paper, we propose replacing the implicit value with an adaptive uncertainty distribution, to differentiate between points based on their distance to the surface. This simple ``value to distribution'' transition yields significant improvements on nearly all the baselines. Furthermore, qualitative results demonstrate that the models trained using our uncertainty distribution loss, can capture more intricate wrinkles, and realistic limbs. Code and models are available for research purposes at https://github.com/psyai-net/D-IF_release

    Reconstruction-Aware Prior Distillation for Semi-supervised Point Cloud Completion

    Full text link
    Point clouds scanned by real-world sensors are always incomplete, irregular, and noisy, making the point cloud completion task become increasingly more important. Though many point cloud completion methods have been proposed, most of them require a large number of paired complete-incomplete point clouds for training, which is labor exhausted. In contrast, this paper proposes a novel Reconstruction-Aware Prior Distillation semi-supervised point cloud completion method named RaPD, which takes advantage of a two-stage training scheme to reduce the dependence on a large-scale paired dataset. In training stage 1, the so-called deep semantic prior is learned from both unpaired complete and unpaired incomplete point clouds using a reconstruction-aware pretraining process. While in training stage 2, we introduce a semi-supervised prior distillation process, where an encoder-decoder-based completion network is trained by distilling the prior into the network utilizing only a small number of paired training samples. A self-supervised completion module is further introduced, excavating the value of a large number of unpaired incomplete point clouds, leading to an increase in the network's performance. Extensive experiments on several widely used datasets demonstrate that RaPD, the first semi-supervised point cloud completion method, achieves superior performance to previous methods on both homologous and heterologous scenarios

    Object Level Depth Reconstruction for Category Level 6D Object Pose Estimation From Monocular RGB Image

    Full text link
    Recently, RGBD-based category-level 6D object pose estimation has achieved promising improvement in performance, however, the requirement of depth information prohibits broader applications. In order to relieve this problem, this paper proposes a novel approach named Object Level Depth reconstruction Network (OLD-Net) taking only RGB images as input for category-level 6D object pose estimation. We propose to directly predict object-level depth from a monocular RGB image by deforming the category-level shape prior into object-level depth and the canonical NOCS representation. Two novel modules named Normalized Global Position Hints (NGPH) and Shape-aware Decoupled Depth Reconstruction (SDDR) module are introduced to learn high fidelity object-level depth and delicate shape representations. At last, the 6D object pose is solved by aligning the predicted canonical representation with the back-projected object-level depth. Extensive experiments on the challenging CAMERA25 and REAL275 datasets indicate that our model, though simple, achieves state-of-the-art performance.Comment: 19 pages, 7 figures, 4 table

    EmoTalk: Speech-Driven Emotional Disentanglement for 3D Face Animation

    Full text link
    Speech-driven 3D face animation aims to generate realistic facial expressions that match the speech content and emotion. However, existing methods often neglect emotional facial expressions or fail to disentangle them from speech content. To address this issue, this paper proposes an end-to-end neural network to disentangle different emotions in speech so as to generate rich 3D facial expressions. Specifically, we introduce the emotion disentangling encoder (EDE) to disentangle the emotion and content in the speech by cross-reconstructed speech signals with different emotion labels. Then an emotion-guided feature fusion decoder is employed to generate a 3D talking face with enhanced emotion. The decoder is driven by the disentangled identity, emotional, and content embeddings so as to generate controllable personal and emotional styles. Finally, considering the scarcity of the 3D emotional talking face data, we resort to the supervision of facial blendshapes, which enables the reconstruction of plausible 3D faces from 2D emotional data, and contribute a large-scale 3D emotional talking face dataset (3D-ETF) to train the network. Our experiments and user studies demonstrate that our approach outperforms state-of-the-art methods and exhibits more diverse facial movements. We recommend watching the supplementary video: https://ziqiaopeng.github.io/emotalkComment: Accepted by ICCV 202
    corecore