145 research outputs found
Identity-adaptive Facial Expression Recognition Through Expression Regeneration Using Conditional Generative Adversarial Networks
Subject variation is a challenging issue for facial expression recognition, especially when handling unseen subjects with small-scale labeled facial expression databases. Although transfer learning has been widely used to tackle the problem, the performance degrades on new data. In this paper, we present a novel approach (so-called IA-gen) to alleviate the issue of subject variations by regenerating expressions from any input facial images. First of all, we train conditional generative models to generate six prototypic facial expressions from any given query face image while keeping the identity related information unchanged. Generative Adversarial Networks are employed to train the conditional generative models, and each of them is designed to generate one of the prototypic facial expression images. Second, a regular CNN (FER-Net) is fine- tuned for expression classification. After the corresponding prototypic facial expressions are regenerated from each facial image, we output the last FC layer of FER-Net as features for both the input image and the generated images. Based on the minimum distance between the input image and the generated expression images in the feature space, the input image is classified as one of the prototypic expressions consequently. Our proposed method can not only alleviate the influence of inter-subject variations but will also be flexible enough to integrate with any other FER CNNs for person-independent facial expression recognition. Our method has been evaluated on CK+, Oulu-CASIA, BU-3DFE and BU-4DFE databases, and the results demonstrate the effectiveness of our proposed method
Ion Transport in the Freshwater Bivalve, Corbicula Fluminea (Muller).
Corbicula fluminea (Muller) maintains hydromineral balance in dilute media by active transepithelial transport of Na and Cl. Cl uptake depended on external (Cl) and displayed saturable kinetics. The transport capacity in pondwater(PW)-acclimated animals was 7.00 0.51 eq(g dry tissue h)\sp{-1} and the transport affinity 0.21 0.08 mM. Prolonged salt depletion increased Cl transport capacity without changing the affinity. In PW-acclimated C. fluminea, Na and Cl transport were independent, and were stimulated by serotonin. In salt depleted (SD)animals, Na transport was partially Cl-dependent but Cl transport was Na-independent. Acetazolamide increased Na and Cl efflux. Both serotonin and acetazolamide promoted the loss of titratable base. N, N\sp\prime-dicyclohexylcarbodiimide (DCCD) inhibited Na and Cl transport. The inhibition of Na transport by DCCD was Cl-dependent, but the inhibition of Cl transport was Na-independent. DCCD increased the loss of titratable base in Cl-free PW medium. 4,4\sp\prime-diisothiocyanostilbene-2,2\sp\prime-disulfonic acid decreased Cl influx and Na net flux. Furosemide inhibited Na and Cl transport in PW-acclimated animals. Exposure of C. fluminea to hyperosmotic nonelectrolytes resulted in an elevation of blood solutes due to dehydration followed by a precipitous decrease in blood Na and the accumulation of nonelectrolytes. Lanthanum was rarely observed to penetrate the paracellular spaces without hyperosmotic stress, but was observed in the paracellular junctional complexes between gill epithelial cells under hyperosmotic conditions. Longer exposure resulted in greater amounts of lanthanum precipitation in more locations compared to shorter exposure. Carunculina texasensis was able to maintain normal blood ion concentration for 8 h with minimal dehydration under hyperosmotic condition. Longer exposure caused a precipitous decrease in most blood solutes and an extensive accumulation of nonelectrolytes. More lanthanum was observed in the paracellular spaces of C. texasensis compared to C. fluminea for identical treatments. This study suggests that Na and Cl transport in C. fluminea are energized by a proton-pump. The entrance of Na is likely via an apical Na-conductive channel and Cl via an apical Cl/HCO\sb3 exchange pathway. It also suggests that under normal conditions the epithelium of C. fluminea is relatively tight and the paracellular pathway for solutes is insignificant
Layer Coherence Origin of Intrinsic Planar Hall Effect in 2D Limit
The intrinsic planar Hall effect has attracted intensive interest inspired by
recent experiments. Existing theories of this effect require three dimensional
orbital motion, or strong spin-orbit coupling of certain forms, which do not
exist in van der Waals thin films. Here, we uncover a new origin of the planar
Hall effect - as an intrinsic property of layer coherent electrons - that
allows its presence even in bilayer and trilayer atomically thin limit. As
examples, we show that the effect can be triggered by strain and interlayer
sliding respectively in twisted bilayer graphene and trilayer transition metal
dichalcogenides, where the effect features rich tunability and even stronger
magnitude than those induced by topological nodal structures in bulk materials.
The layer mechanism also provides a new route towards quantized Hall response
upon a topological phase transition induced by in-plane magnetic field. These
results unveil the unexplored potential of quantum layertronics and moir\'e
flat band for planar Hall transport.Comment: 6 pages, 5 figure
Interlayer electric multipoles induced by in-plane field from quantum geometric origins
We show that interlayer charge transfer in 2D materials can be driven by an
in-plane electric field, giving rise to electrical multipole generation in
linear and second order of in-plane field. The linear and nonlinear effects
have quantum geometric origins in the Berry curvature and quantum metric
respectively, defined in extended parameter spaces characteristic of layered
materials. We elucidate their symmetry characters, and demonstrate sizable
dipole and quadrupole polarizations respectively in twisted bilayers and
trilayers of transition metal dichalcogenides. Furthermore, we show that the
effect is strongly enhanced during the topological phase transition tuned by
interlayer translation. The effects point to a new electric control on layer
quantum degree of freedom.Comment: 13 pages, 4 figure
Multi-modality Empowered Network For Facial Action Unit Detection
This paper presents a new thermal empowered multi-task network (TEMT-Net) to improve facial action unit detection. Our primary goal is to leverage the situation that the training set has multi-modality data while the application scenario only has one modality. Thermal images are robust to illumination and face color. In the proposed multi-task framework, we utilize both modality data. Action unit detection and facial landmark detection are correlated tasks. To utilize the advantage and the correlation of different modalities and different tasks, we propose a novel thermal empowered multi-task deep neural network learning approach for action unit detection, facial landmark detection and thermal image reconstruction simultaneously. The thermal image generator and facial landmark detection provide regularization on the learned features with shared factors as the input color images. Extensive experiments are conducted on the BP4D and MMSE databases, with the comparison to the state-of-the-art methods. The experiments show that the multi-modality framework improves the AU detection significantly
Adaptive Multimodal Fusion For Facial Action Units Recognition
Multimodal facial action units (AU) recognition aims to build models that are capable of processing, correlating, and integrating information from multiple modalities (i.e., 2D images from a visual sensor, 3D geometry from 3D imaging, and thermal images from an infrared sensor). Although the multimodal data can provide rich information, there are two challenges that have to be addressed when learning from multimodal data: 1) the model must capture the complex cross-modal interactions in order to utilize the additional and mutual information effectively; 2) the model must be robust enough in the circumstance of unexpected data corruptions during testing, in case of a certain modality missing or being noisy. In this paper, we propose a novel Adaptive Multimodal Fusion method (AMF) for AU detection, which learns to select the most relevant feature representations from different modalities by a re-sampling procedure conditioned on a feature scoring module. The feature scoring module is designed to allow for evaluating the quality of features learned from multiple modalities. As a result, AMF is able to adaptively select more discriminative features, thus increasing the robustness to missing or corrupted modalities. In addition, to alleviate the over-fitting problem and make the model generalize better on the testing data, a cut-switch multimodal data augmentation method is designed, by which a random block is cut and switched across multiple modalities. We have conducted a thorough investigation on two public multimodal AU datasets, BP4D and BP4D+, and the results demonstrate the effectiveness of the proposed method. Ablation studies on various circumstances also show that our method remains robust to missing or noisy modalities during tests
The 2nd 3D Face Alignment In The Wild Challenge (3DFAW-video): Dense Reconstruction From Video
3D face alignment approaches have strong advantages over 2D with respect to representational power and robustness to illumination and pose. Over the past few years, a number of research groups have made rapid advances in dense 3D alignment from 2D video and obtained impressive results. How these various methods compare is relatively unknown. Previous benchmarks addressed sparse 3D alignment and single image 3D reconstruction. No commonly accepted evaluation protocol exists for dense 3D face reconstruction from video with which to compare them. The 2nd 3D Face Alignment in the Wild from Videos (3DFAW-Video) Challenge extends the previous 3DFAW 2016 competition to the estimation of dense 3D facial structure from video. It presented a new large corpora of profile-to-profile face videos recorded under different imaging conditions and annotated with corresponding high-resolution 3D ground truth meshes. In this paper we outline the evaluation protocol, the data used, and the results. 3DFAW-Video is to be held in conjunction with the 2019 International Conference on Computer Vision, in Seoul, Korea
EmbeddingTree: Hierarchical Exploration of Entity Features in Embedding
Embedding learning transforms discrete data entities into continuous
numerical representations, encoding features/properties of the entities.
Despite the outstanding performance reported from different embedding learning
algorithms, few efforts were devoted to structurally interpreting how features
are encoded in the learned embedding space. This work proposes EmbeddingTree, a
hierarchical embedding exploration algorithm that relates the semantics of
entity features with the less-interpretable embedding vectors. An interactive
visualization tool is also developed based on EmbeddingTree to explore
high-dimensional embeddings. The tool helps users discover nuance features of
data entities, perform feature denoising/injecting in embedding training, and
generate embeddings for unseen entities. We demonstrate the efficacy of
EmbeddingTree and our visualization tool through embeddings generated for
industry-scale merchant data and the public 30Music listening/playlists
dataset.Comment: 5 pages, 3 figures, accepted by PacificVis 202
- …