68,784 research outputs found

    Are Face and Object Recognition Independent? A Neurocomputational Modeling Exploration

    Full text link
    Are face and object recognition abilities independent? Although it is commonly believed that they are, Gauthier et al.(2014) recently showed that these abilities become more correlated as experience with nonface categories increases. They argued that there is a single underlying visual ability, v, that is expressed in performance with both face and nonface categories as experience grows. Using the Cambridge Face Memory Test and the Vanderbilt Expertise Test, they showed that the shared variance between Cambridge Face Memory Test and Vanderbilt Expertise Test performance increases monotonically as experience increases. Here, we address why a shared resource across different visual domains does not lead to competition and to an inverse correlation in abilities? We explain this conundrum using our neurocomputational model of face and object processing (The Model, TM). Our results show that, as in the behavioral data, the correlation between subordinate level face and object recognition accuracy increases as experience grows. We suggest that different domains do not compete for resources because the relevant features are shared between faces and objects. The essential power of experience is to generate a "spreading transform" for faces that generalizes to objects that must be individuated. Interestingly, when the task of the network is basic level categorization, no increase in the correlation between domains is observed. Hence, our model predicts that it is the type of experience that matters and that the source of the correlation is in the fusiform face area, rather than in cortical areas that subserve basic level categorization. This result is consistent with our previous modeling elucidating why the FFA is recruited for novel domains of expertise (Tong et al., 2008)

    The new accent technologies:recognition, measurement and manipulation of accented speech

    Get PDF

    Cognitive processes in categorical and associative priming: a diffusion model analysis

    Get PDF
    Cognitive processes and mechanisms underlying different forms of priming were investigated using a diffusion model approach. In a series of 6 experiments, effects of prime-target associations and of a semantic and affective categorical match of prime and target were analyzed for different tasks. Significant associative and categorical priming effects were found in standard analyses of response times (RTs) and error frequencies. Results of diffusion model analyses revealed that priming effects of associated primes were mapped on the drift rate parameter (v), while priming effects of a categorical match on a task-relevant dimension were mapped on the extradecisional parameters (t(0) and d). These results support a spreading activation account of associative priming and an explanation of categorical priming in terms of response competition. Implications for the interpretation of priming effects and the use of priming paradigms in cognitive psychology and social cognition are discussed

    A New Vehicle Localization Scheme Based on Combined Optical Camera Communication and Photogrammetry

    Full text link
    The demand for autonomous vehicles is increasing gradually owing to their enormous potential benefits. However, several challenges, such as vehicle localization, are involved in the development of autonomous vehicles. A simple and secure algorithm for vehicle positioning is proposed herein without massively modifying the existing transportation infrastructure. For vehicle localization, vehicles on the road are classified into two categories: host vehicles (HVs) are the ones used to estimate other vehicles' positions and forwarding vehicles (FVs) are the ones that move in front of the HVs. The FV transmits modulated data from the tail (or back) light, and the camera of the HV receives that signal using optical camera communication (OCC). In addition, the streetlight (SL) data are considered to ensure the position accuracy of the HV. Determining the HV position minimizes the relative position variation between the HV and FV. Using photogrammetry, the distance between FV or SL and the camera of the HV is calculated by measuring the occupied image area on the image sensor. Comparing the change in distance between HV and SLs with the change in distance between HV and FV, the positions of FVs are determined. The performance of the proposed technique is analyzed, and the results indicate a significant improvement in performance. The experimental distance measurement validated the feasibility of the proposed scheme

    DolphinAtack: Inaudible Voice Commands

    Full text link
    Speech recognition (SR) systems such as Siri or Google Now have become an increasingly popular human-computer interaction method, and have turned various systems into voice controllable systems(VCS). Prior work on attacking VCS shows that the hidden voice commands that are incomprehensible to people can control the systems. Hidden voice commands, though hidden, are nonetheless audible. In this work, we design a completely inaudible attack, DolphinAttack, that modulates voice commands on ultrasonic carriers (e.g., f > 20 kHz) to achieve inaudibility. By leveraging the nonlinearity of the microphone circuits, the modulated low frequency audio commands can be successfully demodulated, recovered, and more importantly interpreted by the speech recognition systems. We validate DolphinAttack on popular speech recognition systems, including Siri, Google Now, Samsung S Voice, Huawei HiVoice, Cortana and Alexa. By injecting a sequence of inaudible voice commands, we show a few proof-of-concept attacks, which include activating Siri to initiate a FaceTime call on iPhone, activating Google Now to switch the phone to the airplane mode, and even manipulating the navigation system in an Audi automobile. We propose hardware and software defense solutions. We validate that it is feasible to detect DolphinAttack by classifying the audios using supported vector machine (SVM), and suggest to re-design voice controllable systems to be resilient to inaudible voice command attacks.Comment: 15 pages, 17 figure

    Further evidence on the effect of symbolic distance on Stroop-like interference

    Get PDF
    Pavese and UmiltĂ  found that, in an enumeration task, Stroop-like interference is larger when the digit identity is symbolically close to the enumeration response than when it is symbolically far. In two experiments testing 49 undergraduates, we further explored this phenomenon using Francolini and Egeth's paradigm. We found that symbolic distance affected interference even when the stimulus was briefly presented and masked. In Exp. 2, which tested numerosities outside the subitizing range, individuals used a different enumeration strategy but showed the same symbolic distance effect. These results support the hypothesis that Stroop interference found in enumeration tasks depends on a rapid and automatic activation of digits' magnitude representation
    • …
    corecore