20 research outputs found

    Contrast Enhancement of Brightness-Distorted Images by Improved Adaptive Gamma Correction

    Full text link
    As an efficient image contrast enhancement (CE) tool, adaptive gamma correction (AGC) was previously proposed by relating gamma parameter with cumulative distribution function (CDF) of the pixel gray levels within an image. ACG deals well with most dimmed images, but fails for globally bright images and the dimmed images with local bright regions. Such two categories of brightness-distorted images are universal in real scenarios, such as improper exposure and white object regions. In order to attenuate such deficiencies, here we propose an improved AGC algorithm. The novel strategy of negative images is used to realize CE of the bright images, and the gamma correction modulated by truncated CDF is employed to enhance the dimmed ones. As such, local over-enhancement and structure distortion can be alleviated. Both qualitative and quantitative experimental results show that our proposed method yields consistently good CE results

    Deactivation Effects of Tb3+ on Ho3+ Emission in Fluoroindate Glasses for 3.9 μm Laser Applications

    Get PDF
    A series of Ho3+/Tb3+ co-doped fluoroindate glasses with good thermal stability have been synthesized to study the deactivation effects of Tb3+ on the Ho3+: 3.9 μm emission. Efficient 3.9 μm emission enhancement is obtained under excitation by an 888 nm laser diode (LD). The Judd-Ofelt (J-O) intensity parameters and radiative properties are calculated to evaluate the spectroscopic properties. Possible energy transfer processes resulting in emission reinforcement are discussed. A higher spontaneous transition probability and larger peak emission cross section are achieved with the inclusion of Tb3+. This analysis supports the conclusion that Ho3+/Tb3+ co-doped fluoroindate glass is a potentially useful laser material for highly efficient 3.9 μm fiber lasers

    PMMTalk: Speech-Driven 3D Facial Animation from Complementary Pseudo Multi-modal Features

    Full text link
    Speech-driven 3D facial animation has improved a lot recently while most related works only utilize acoustic modality and neglect the influence of visual and textual cues, leading to unsatisfactory results in terms of precision and coherence. We argue that visual and textual cues are not trivial information. Therefore, we present a novel framework, namely PMMTalk, using complementary Pseudo Multi-Modal features for improving the accuracy of facial animation. The framework entails three modules: PMMTalk encoder, cross-modal alignment module, and PMMTalk decoder. Specifically, the PMMTalk encoder employs the off-the-shelf talking head generation architecture and speech recognition technology to extract visual and textual information from speech, respectively. Subsequently, the cross-modal alignment module aligns the audio-image-text features at temporal and semantic levels. Then PMMTalk decoder is employed to predict lip-syncing facial blendshape coefficients. Contrary to prior methods, PMMTalk only requires an additional random reference face image but yields more accurate results. Additionally, it is artist-friendly as it seamlessly integrates into standard animation production workflows by introducing facial blendshape coefficients. Finally, given the scarcity of 3D talking face datasets, we introduce a large-scale 3D Chinese Audio-Visual Facial Animation (3D-CAVFA) dataset. Extensive experiments and user studies show that our approach outperforms the state of the art. We recommend watching the supplementary video

    Surface acoustic wave‐enhanced multi‐view acoustofluidic rotation cytometry (MARC) for pre‐cytopathological screening

    Get PDF
    Cytopathology, crucial in disease diagnosis, commonly uses microscopic slides to scrutinize cellular abnormalities. However, processing high volumes of samples often results in numerous negative diagnoses, consuming significant time and resources in healthcare. To address this challenge, a surface acoustic wave‐enhanced multi‐view acoustofluidic rotation cytometry (MARC) technique is developed for pre‐cytopathological screening. MARC enhances cellular morphology analysis through comprehensive and multi‐angle observations and amplifies subtle cell differences, particularly in the nuclear‐to‐cytoplasmic ratio, across various cell types and between cancerous and normal tissue cells. By prioritizing MARC‐screened positive cases, this approach can potentially streamline traditional cytopathology, reducing the workload and resources spent on negative diagnoses. This significant advancement enhances overall diagnostic efficiency, offering a transformative vision for cytopathological screening

    A Framework for the Multi-Level Fusion of Electronic Nose and Electronic Tongue for Tea Quality Assessment

    No full text
    Electronic nose (E-nose) and electronic tongue (E-tongue) can mimic the sensory perception of human smell and taste, and they are widely applied in tea quality evaluation by utilizing the fingerprints of response signals representing the overall information of tea samples. The intrinsic part of human perception is the fusion of sensors, as more information is provided comparing to the information from a single sensory organ. In this study, a framework for a multi-level fusion strategy of electronic nose and electronic tongue was proposed to enhance the tea quality prediction accuracies, by simultaneously modeling feature fusion and decision fusion. The procedure included feature-level fusion (fuse the time-domain based feature and frequency-domain based feature) and decision-level fusion (D-S evidence to combine the classification results from multiple classifiers). The experiments were conducted on tea samples collected from various tea providers with four grades. The large quantity made the quality assessment task very difficult, and the experimental results showed much better classification ability for the multi-level fusion system. The proposed algorithm could better represent the overall characteristics of tea samples for both odor and taste

    Ensemble feature learning for material recognition with convolutional neural networks

    No full text
    Abstract Material recognition is the process of recognizing the constituent material of the object, and it is a crucial step in many fields. Therefore, it is valuable to create a system that could achieve material recognition automatically. This paper proposes a novel approach named ensemble learning for material recognition with convolutional neural networks (CNNs). In the proposed method, firstly, a CNN model is trained to extract the image features. Secondly, knowledge-based classifiers are learned to get the probabilities of the test sample that belongs to different material categories. Finally, we propose three different ways to learn the ensemble features, which achieves higher recognition accuracy. The great difference from the prior work is that we combine the knowledge-based classifiers on probability level. Experimental results show that the proposed ensemble feature learning method performs better than the state-of-the-art material recognition methods and can archive a much higher recognition accuracy

    Infants\u27 Pain Recognition Based on Facial Expression: Dynamic Hybrid Descriptions

    No full text
    Abstract The accurate assessment of infants\u27 pain is important for understanding their medical conditions and developing suitable treatment. Pediatric studies reported that the inadequate treatment of infants\u27 pain might cause various neuroanatomical and psychological problems. The fact that infants can not communicate verbally motivates increasing interests to develop automatic pain assessment system that provides continuous and accurate pain assessment. In this paper, we propose a new set of pain facial activity features to describe the infants\u27 facial expression of pain. Both dynamic facial texture feature and dynamic geometric feature are extracted from video sequences and utilized to classify facial expression of infants as pain or no pain. For the dynamic analysis of facial expression, we construct spatiotemporal domain representation for texture features and time series representation (i.e. time series of frame-level features) for geometric features. Multiple facial features are combined through both feature fusion and decision fusion schemes to evaluate their effectiveness in infants\u27 pain assessment. Experiments are conducted on the video acquired from NICU infants, and the best accuracy of the proposed pain assessment approaches is 95.6%. Moreover, we find that although decision fusion does not perform better than that of feature fusion, the False Negative Rate of decision fusion (6.2%) is much lower than that of feature fusion (25%)

    Automatic Infants’ Pain Assessment by Dynamic Facial Representation: Effects of Profile View, Gestational Age, Gender, and Race

    No full text
    Infants’ early exposure to painful procedures can have negative short and long-term effects on cognitive, neurological, and brain development. However, infants cannot express their subjective pain experience, as they do not communicate in any language. Facial expression is the most specific pain indicator, which has been effectively employed for automatic pain recognition. In this paper, dynamic pain facial expression representation and fusion scheme for automatic pain assessment in infants is proposed by combining temporal appearance facial features and temporal geometric facial features. We investigate the effects of various factors that influence pain reactivity in infants, such as individual variables of gestational age, gender, and race. Different automatic infant pain assessment models are constructed, depending on influence factors as well as facial profile view, which affect the model ability of pain recognition. It can be concluded that the profile-based infant pain assessment is feasible, as its performance is almost as good as that of the whole face. Moreover, gestational age is the most influencing factor for pain assessment, and it is necessary to construct specific models depending on it. This is mainly because of a lack of behavioral communication ability in infants with low gestational age, due to limited neurological development. To our best knowledge, this is the first study investigating infants’ pain recognition, highlighting profile facial views and various individual variables

    Infants' Pain Recognition Based on Facial Expression: Dynamic Hybrid Descriptions

    No full text
    corecore