94 research outputs found

    Fast and Interpretable Nonlocal Neural Networks for Image Denoising via Group-Sparse Convolutional Dictionary Learning

    Full text link
    Nonlocal self-similarity within natural images has become an increasingly popular prior in deep-learning models. Despite their successful image restoration performance, such models remain largely uninterpretable due to their black-box construction. Our previous studies have shown that interpretable construction of a fully convolutional denoiser (CDLNet), with performance on par with state-of-the-art black-box counterparts, is achievable by unrolling a dictionary learning algorithm. In this manuscript, we seek an interpretable construction of a convolutional network with a nonlocal self-similarity prior that performs on par with black-box nonlocal models. We show that such an architecture can be effectively achieved by upgrading the 1\ell 1 sparsity prior of CDLNet to a weighted group-sparsity prior. From this formulation, we propose a novel sliding-window nonlocal operation, enabled by sparse array arithmetic. In addition to competitive performance with black-box nonlocal DNNs, we demonstrate the proposed sliding-window sparse attention enables inference speeds greater than an order of magnitude faster than its competitors.Comment: 11 pages, 8 figures, 6 table

    Lateralization in the dichotic listening of tones is influenced by the content of speech

    Get PDF
    Available online 10 February 2020.Cognitive functions, for example speech processing, are distributed asymmetrically in the two hemispheres that mostly have homologous anatomical structures. Dichotic listening is a well-established paradigm to investigate hemispherical lateralization of speech. However, the mixed results of dichotic listening, especially when using tonal languages as stimuli, complicates the investigation of functional lateralization. We hypothesized that the inconsistent results in dichotic listening are due to an interaction in processing a mixture of acoustic and linguistic attributes that are differentially processed over the two hemispheres. In this study, a within-subject dichotic listening paradigm was designed, in which different levels of speech and linguistic information was incrementally included in different conditions that required the same tone identification task. A left ear advantage (LEA), in contrast with the commonly found right ear advantage (REA) in dichotic listening, was observed in the hummed tones condition, where only the slow frequency modulation of tones was included. However, when phonemic and lexical information was added in simple vowel tone conditions, the LEA became unstable. Furthermore, ear preference became balanced when phonological and lexicalsemantic attributes were included in the consonant-vowel (CV), pseudo-word, and word conditions. Compared with the existing REA results that use complex vowel word tones, a complete pattern emerged gradually shifting from LEA to REA. These results support the hypothesis that an acoustic analysis of suprasegmental information of tones is preferably processed in the right hemisphere, but is influenced by phonological and lexical semantic processes residing in the left hemisphere. The ear preference in dichotic listening depends on the levels of speech and linguistic analysis and preferentially lateralizes across the different hemispheres. That is, the manifestation of functional lateralization depends on the integration of information across the two hemispheres.This study was supported by National Natural Science Foundation of China 31871131, Major Program of Science and Technology Commission of Shanghai Municipality (STCSM) 17JC1404104, Program of Introducing Talents of Discipline to Universities, Base B16018 to XT, and the JRI Seed Grants for Research Collaboration from NYU-ECNU Institute of Brain and Cognitive Science at NYU Shanghai to XT and QC, and NIH 2R01DC05660 to David Poeppel at New York University supporting NM and AF and F32 DC011985 to AF

    Bridging the Gap Between Big Data and Social Services

    No full text
    The goal of health and human service agencies is to benefit the general public as well as protect at-risk populations from worsening social concerns. While there has been a growing focus on prevention, predictive models can be hard to translate into solutions that can be effectively implemented. The recent proliferation of big data sources has created an unprecedented opportunity to leverage data in order to focus work with vulnerable populations and provide predictive-based intervention prior to the worsening of an individual’s situation. For example, publically available court records indicating an imminent eviction can be used in order to identify a population at a greater risk of becoming homeless. Prevention services can be provided to these identified individuals prior to their becoming homeless. This intervention, which precedes actual homelessness, not only helps an individual or family, but is also cost effective for the city. Such an approach requires integrating solutions across multiple levels: data integrity, predictive analytics, and implementing an effective intervention process. There are not many organizations that have the necessary tools, ability and knowledge to follow through on all these levels in order to deliver an effective outcome. In this perspective we would like to introduce a predictive-based social intervention approach and examine the associated challenges that must be addressed
    corecore