77 research outputs found

    A Zero-/Few-Shot Anomaly Classification and Segmentation Method for CVPR 2023 VAND Workshop Challenge Tracks 1&2: 1st Place on Zero-shot AD and 4th Place on Few-shot AD

    Full text link
    In this technical report, we briefly introduce our solution for the Zero/Few-shot Track of the Visual Anomaly and Novelty Detection (VAND) 2023 Challenge. For industrial visual inspection, building a single model that can be rapidly adapted to numerous categories without or with only a few normal reference images is a promising research direction. This is primarily because of the vast variety of the product types. For the zero-shot track, we propose a solution based on the CLIP model by adding extra linear layers. These layers are used to map the image features to the joint embedding space, so that they can compare with the text features to generate the anomaly maps. Besides, when the reference images are available, we utilize multiple memory banks to store their features and compare them with the features of the test images during the testing phase. In this challenge, our method achieved first place in the zero-shot track, especially excelling in segmentation with an impressive F1 score improvement of 0.0489 over the second-ranked participant. Furthermore, in the few-shot track, we secured the fourth position overall, with our classification F1 score of 0.8687 ranking first among all participating teams

    MindShift: Leveraging Large Language Models for Mental-States-Based Problematic Smartphone Use Intervention

    Full text link
    Problematic smartphone use negatively affects physical and mental health. Despite the wide range of prior research, existing persuasive techniques are not flexible enough to provide dynamic persuasion content based on users' physical contexts and mental states. We first conduct a Wizard-of-Oz study (N=12) and an interview study (N=10) to summarize the mental states behind problematic smartphone use: boredom, stress, and inertia. This informs our design of four persuasion strategies: understanding, comforting, evoking, and scaffolding habits. We leverage large language models (LLMs) to enable the automatic and dynamic generation of effective persuasion content. We develop MindShift, a novel LLM-powered problematic smartphone use intervention technique. MindShift takes users' in-the-moment physical contexts, mental states, app usage behaviors, users' goals & habits as input, and generates high-quality and flexible persuasive content with appropriate persuasion strategies. We conduct a 5-week field experiment (N=25) to compare MindShift with baseline techniques. The results show that MindShift significantly improves intervention acceptance rates by 17.8-22.5% and reduces smartphone use frequency by 12.1-14.4%. Moreover, users have a significant drop in smartphone addiction scale scores and a rise in self-efficacy. Our study sheds light on the potential of leveraging LLMs for context-aware persuasion in other behavior change domains

    Enzyme-Free Electrochemical Glucose Sensors Prepared by Dealloying Pd-Ni-P Metallic Glasses

    Get PDF
    We report the formation of enzyme-free electrochemical glucose sensors by electrochemical dealloying palladium-containing Pd-Ni-P metallic glasses. When metallic glasses with different Pd contents are used as the dealloying precursor alloys, palladium-based nanoporous metals with different ligament and pore sizes can be obtained. The chemical compositions of the nanoporous metals also vary according to the different precursor compositions. All the as-obtained nanoporous metals exhibit electrochemical catalytic activity towards the oxidation of d-glucose, indicating that the nanoporous metals prepared by dealloying the Pd-Ni-P metallic glasses are promising materials for enzyme-free electrochemical glucose sensor

    The Contribution of Sound Intensity in Vocal Emotion Perception: Behavioral and Electrophysiological Evidence

    Get PDF
    Although its role is frequently stressed in acoustic profile for vocal emotion, sound intensity is frequently regarded as a control parameter in neurocognitive studies of vocal emotion, leaving its role and neural underpinnings unclear. To investigate these issues, we asked participants to rate the angry level of neutral and angry prosodies before and after sound intensity modification in Experiment 1, and recorded electroencephalogram (EEG) for mismatching emotional prosodies with and without sound intensity modification and for matching emotional prosodies while participants performed emotional feature or sound intensity congruity judgment in Experiment 2. It was found that sound intensity modification had significant effect on the rating of angry level for angry prosodies, but not for neutral ones. Moreover, mismatching emotional prosodies, relative to matching ones, induced enhanced N2/P3 complex and theta band synchronization irrespective of sound intensity modification and task demands. However, mismatching emotional prosodies with reduced sound intensity showed prolonged peak latency and decreased amplitude in N2/P3 complex and smaller theta band synchronization. These findings suggest that though it cannot categorically affect emotionality conveyed in emotional prosodies, sound intensity contributes to emotional significance quantitatively, implying that sound intensity should not simply be taken as a control parameter and its unique role needs to be specified in vocal emotion studies

    Effects of Aging Stereotype Threat on Working Self-Concepts: An Event-Related Potentials Approach

    Get PDF
    Although the influence of stereotype threat (ST) on working self-concepts has been highlighted in recent years, its neural underpinnings are unclear. Notably, the aging ST, which largely influences older adults’ cognitive ability, mental and physical health, did not receive much attention. In order to investigate these issues, electroencephalogram (EEG) data were obtained from older adults during a modified Stroop task using neutral words, positive and negative self-concept words in aging ST vs. neutral control conditions. Results showed longer reaction times (RTs) for identifying colors of words under the aging ST compared to the neutral condition. More importantly, the negative self-concept elicited more positive late P300 amplitudes and enhanced theta band activities compared to the positive self-concept or neutral words under the aging ST condition, whereas no difference was found between these self-concepts and neutral words in the control condition. Furthermore, the aging ST induced smaller theta band synchronization and enhanced alpha band synchronization compared to the control condition. Moreover, we also observed valence differences in self-concepts where the negative self-concept words reduced early P150/N170 complex relative to neutral words. These findings suggest that priming ST could activate negative self-concepts as current working self-concept, and that this influence occurred during a late neural time course

    When Brain Differentiates Happy from Neutral in Prosody?

    No full text
    The effect of different intensities of vocal emotion on event related potentials has yet not been studied. We therefore investigated 16 healthy participants with emotion and sound decision on neutral and happy voice which varied continuously in intensity. The result found that neutral and happy voice can be differentiated on P2 component under both explicit and implicit condition. Moreover, the P2 parameters were linear correlated with the rate of happiness, suggesting a graded processing of vocal emotion in early stage. However, the brain distinguished neutral from happy in P3 interval when performing explicit task but exhibit a categorical feature.</p

    How Distance Affects Semantic Integration in Discourse: Evidence from Event-Related Potentials.

    No full text
    Event-related potentials were used to investigate whether semantic integration in discourse is influenced by the number of intervening sentences between the endpoints of integration. Readers read discourses in which the last sentence contained a critical word that was either congruent or incongruent with the information introduced in the first sentence. Furthermore, for the short discourses, the first and last sentence were intervened by only one sentence while for the long discourses, they were intervened by three sentences. We found that the incongruent words elicited an N400 effect for both the short and long discourses. However, a P600 effect was only observed for the long discourses, but not for the short ones. These results suggest that although readers can successfully integrate upcoming words into the existing discourse representation, the effort required for this integration process is modulated by the number of intervening sentences. Thus, discourse distance as measured by the number of intervening sentences should be taken as an important factor for semantic integration in discourse

    Event-related potential correlates of the expectancy violation effect during emotional prosody processing

    No full text
    The present study investigated the expectancy violation effects evoked by deviation in sentential emotional prosody (EP), and their association with the deviation patterns. Event-related potentials (ERPs) were recorded for mismatching EPs with different patterns of deviation and for matching control EPs while subjects performed emotional congruousness judgment in Experiment 1 and visual probe detection tasks in Experiment 2. In the control experiment, EPs and acoustically matched non-emotional materials were presented and ERPs were recorded while participants judged the sound intensity congruousness. It was found that an early negativity, whose peak latency varied with deviation pattern, was elicited by mismatching EPs relative to matching ones, irrespective of task-relevance. A late positivity was specifically induced by mismatching EPs, and was modulated by both deviation pattern and task-relevance. Moreover, these effects cannot be simply attributed to the change in non-emotional acoustic properties. These findings suggest that the brain detects the EP deviation rapidly, and then integrates it with context for comprehension, during which the emotionality plays a role of speeding up the perception and enhancing vigilance. (C) 2010 Elsevier B.V. All rights reserved

    Emotional Intensity Modulates the Integration of Bimodal Angry Expressions: ERP Evidence

    No full text
    Integration of information from face and voice plays a central role in social interactions. The present study investigated the modulation of emotional intensity on the integration of facial-vocal emotional cues by recording EEG for participants while they were performing emotion identification task on facial, vocal, and bimodal angry expressions varying in emotional intensity. Behavioral results showed the rates of anger and reaction speed increased as emotional intensity across modalities. Critically, the P2 amplitudes were larger for bimodal expressions than for the sum of facial and vocal expressions for low emotional intensity stimuli, but not for middle and high emotional intensity stimuli. These findings suggested that emotional intensity modulates the integration of facial-vocal angry expressions, following the principle of Inverse Effectiveness (IE) in multimodal sensory integration
    corecore