3,146 research outputs found

    Proactive listening to a training commentary improves hazard prediction

    Get PDF
    The aim of this work was to explore the effect of Proactive Listening to a Training Commentary, using the recently developed version of the Spanish Hazard Perception test. Firstly, 16 videos were used in the pre-test session in its short version, cut to black just before the hazard appearance. The What Happens Next Assessment (at the pre-test stage) generates expectations about the outcome of the traffic situation. Then, the training (8 minutes in length) uses the complete version of the same 16 videos, revealing the hazards unfolding. It involves listening to a voice with relevant information about where to allocate attention in the complex driving scene in order to recognise and anticipate the hazard successfully. A total of 121 participants were included in this study The sample consisted of learner, novice and experienced drivers, including re-offender and non-offender drivers. The participants were divided into 2 groups: a trained and an untrained group. Two assessment times were used: pre-test (16 videos) and post-test sessions (another 16 videos). The test presented a high internal consistency (Alpha = 0.875). This training shows significant positive effects for all types and groups of participants. No significant differences were found between the non-offender and the offender groups. Performance in gradual-onset hazard events can be improved after training but also by practice; however this training is essential and especially beneficial for training the ability to detect hazards that appear abruptly (which seems to be difficult to improve just by practice)

    Tencent AVS: A Holistic Ads Video Dataset for Multi-modal Scene Segmentation

    Full text link
    Temporal video segmentation and classification have been advanced greatly by public benchmarks in recent years. However, such research still mainly focuses on human actions, failing to describe videos in a holistic view. In addition, previous research tends to pay much attention to visual information yet ignores the multi-modal nature of videos. To fill this gap, we construct the Tencent `Ads Video Segmentation'~(TAVS) dataset in the ads domain to escalate multi-modal video analysis to a new level. TAVS describes videos from three independent perspectives as `presentation form', `place', and `style', and contains rich multi-modal information such as video, audio, and text. TAVS is organized hierarchically in semantic aspects for comprehensive temporal video segmentation with three levels of categories for multi-label classification, e.g., `place' - `working place' - `office'. Therefore, TAVS is distinguished from previous temporal segmentation datasets due to its multi-modal information, holistic view of categories, and hierarchical granularities. It includes 12,000 videos, 82 classes, 33,900 segments, 121,100 shots, and 168,500 labels. Accompanied with TAVS, we also present a strong multi-modal video segmentation baseline coupled with multi-label class prediction. Extensive experiments are conducted to evaluate our proposed method as well as existing representative methods to reveal key challenges of our dataset TAVS

    Digital Image Access & Retrieval

    Get PDF
    The 33th Annual Clinic on Library Applications of Data Processing, held at the University of Illinois at Urbana-Champaign in March of 1996, addressed the theme of "Digital Image Access & Retrieval." The papers from this conference cover a wide range of topics concerning digital imaging technology for visual resource collections. Papers covered three general areas: (1) systems, planning, and implementation; (2) automatic and semi-automatic indexing; and (3) preservation with the bulk of the conference focusing on indexing and retrieval.published or submitted for publicatio

    Current Resources

    Get PDF
    Reviews include: Family Centered Services: A Handbook for Practitioners.Bonnie K. Williams (Ed.). The National Resource Center for Family Centered Practice, School of Social Work, The University of Iowa. Iowa City, Iowa.Reviewed by Lois Wright Building Skills in High-Risk Families: Strategies for the Home-Based Practitioner. Jane Peterson, Paula E. Kohrt, Linda M. Shadoin, Karen J. Authier. Boys Town, Nebraska. Boys Town Press. Reviewed by Sharon Alper

    Changes in lifeguards’ hazard detection and eye movements with experience: is one season enough?

    Get PDF
    Surveillance is key to the lifesaving capability of lifeguards. Experienced personnel consistently display enhanced hazard detection capabilities compared to less experienced counterparts. However, the mechanisms which underpin this effect and the time it takes to develop these skills are not understood. We hypothesized that, after one season of experience, the number of hazards detected by, and eye movements of, less experienced lifeguards (LEL) would more closely approximate experienced lifeguards (EL). The LEL watched ‘beach scene’ videos at the beginning and end of their first season. The number of hazards detected and eye-movement data were collected and compared to the EL group. The LEL perceived fewer hazards than EL and did not increase over the season. There was no difference in eye-movements between groups. Findings suggest one season is not enough for lifeguards to develop enhanced hazard detection skills and skill level differences are not underpinned by differences in gaze behavior

    Changes in Lifeguards’ Hazard Detection and Eye Movements with Experience: Is One Season Enough?

    Get PDF
    Surveillance is key to the lifesaving capability of lifeguards. Experienced personnel consistently display enhanced hazard detection capabilities compared to less experienced counterparts. However, the mechanisms which underpin this effect and the time it takes to develop these skills are not understood. We hypothesized that, after one season of experience, the number of hazards detected by, and eye movements of, less experienced lifeguards (LEL) would more closely approximate experienced lifeguards (EL). The LEL watched ‘beach scene’ videos at the beginning and end of their first season. The number of hazards detected and eye-movement data were collected and compared to the EL group. The LEL perceived fewer hazards than EL and did not increase over the season. There was no difference in eye-movements between groups. Findings suggest one season is not enough for lifeguards to develop enhanced hazard detection skills and skill level differences are not underpinned by differences in gaze behavior

    "You Tube and I Find" - personalizing multimedia content access

    Full text link
    Recent growth in broadband access and proliferation of small personal devices that capture images and videos has led to explosive growth of multimedia content available everywhereVfrom personal disks to the Web. While digital media capture and upload has become nearly universal with newer device technology, there is still a need for better tools and technologies to search large collections of multimedia data and to find and deliver the right content to a user according to her current needs and preferences. A renewed focus on the subjective dimension in the multimedia lifecycle, fromcreation, distribution, to delivery and consumption, is required to address this need beyond what is feasible today. Integration of the subjective aspects of the media itselfVits affective, perceptual, and physiological potential (both intended and achieved), together with those of the users themselves will allow for personalizing the content access, beyond today’s facility. This integration, transforming the traditional multimedia information retrieval (MIR) indexes to more effectively answer specific user needs, will allow a richer degree of personalization predicated on user intention and mode of interaction, relationship to the producer, content of the media, and their history and lifestyle. In this paper, we identify the challenges in achieving this integration, current approaches to interpreting content creation processes, to user modelling and profiling, and to personalized content selection, and we detail future directions. The structure of the paper is as follows: In Section I, we introduce the problem and present some definitions. In Section II, we present a review of the aspects of personalized content and current approaches for the same. Section III discusses the problem of obtaining metadata that is required for personalized media creation and present eMediate as a case study of an integrated media capture environment. Section IV presents the MAGIC system as a case study of capturing effective descriptive data and putting users first in distributed learning delivery. The aspects of modelling the user are presented as a case study in using user’s personality as a way to personalize summaries in Section V. Finally, Section VI concludes the paper with a discussion on the emerging challenges and the open problems

    Smart augmented reality instructional system for mechanical assembly

    Get PDF
    Quality and efficiency are pivotal indicators of a manufacturing company. Many companies are suffering from shortage of experienced workers across the production line to perform complex assembly tasks such as assembly of an aircraft engine. This could lead to a significant financial loss. In order to further reduce time and error in an assembly, a smart system consisting of multi-modal Augmented Reality (AR) instructions with the support of a deep learning network for tool detection is introduced. The multi-modal smart AR is designed to provide on-site information including various visual renderings with a fine-tuned Region-based Convolutional Neural Network, which is trained on a synthetic tool dataset. The dataset is generated using CAD models of tools augmented onto a 2D scene without the need of manually preparing real tool images. By implementing the system to mechanical assembly of a CNC carving machine, the result has shown that the system is not only able to correctly classify and localize the physical tools but also enables workers to successfully complete the given assembly tasks. With the proposed approaches, an efficiently customizable smart AR instructional system capable of sensing, characterizing the requirements, and enhancing worker\u27s performance effectively has been built and demonstrated --Abstract, page iii
    • …
    corecore