628,342 research outputs found

    Tip of the Iceberg: Part 2, Discovering What\u27s Hidden

    Get PDF
    Remote storage for large collections is becoming common, making those books inaccessible for physical browsing by researchers. The main libraries at Temple University and the University of Central Florida (UCF) each have approximately 1.3 million print items on-site. Both libraries are storing 90% of their collections in automated retrieval systems with 10% remaining available for browsing in open stacks. In Part 1, “Choosing What Shows,” Karen Kohn, Temple’s Collection Analysis Librarian, describes the decisions and processes used for the 10% left physically visible. This second part explores UCF’s efforts to improve discoverability of the items in storage. The visual aspects of a book (height, multi-volume, etc.) that often provide useful clues regarding the content disappear when the patron can only view a list of search results on a computer screen. How can the loss of these visual clues be mitigated? Online browsing guides have been created at UCF to help researchers explore the hierarchical subject structure of call numbers. Other finding tools and displays may also improve awareness of stored materials

    Feature-based tracking of multiple people for intelligent video surveillance.

    Get PDF
    Intelligent video surveillance is the process of performing surveillance task automatically by a computer vision system. It involves detecting and tracking people in the video sequence and understanding their behavior. This thesis addresses the problem of detecting and tracking multiple moving people with unknown background. We have proposed a feature-based framework for tracking, which requires feature extraction and feature matching. We have considered color, size, blob bounding box and motion information as features of people. In our feature-based tracking system, we have proposed to use Pearson correlation coefficient for matching feature-vector with temporal templates. The occlusion problem has been solved by histogram backprojection. Our tracking system is fast and free from assumptions about human structure. We have implemented our tracking system using Visual C++ and OpenCV and tested on real-world images and videos. Experimental results suggest that our tracking system achieved good accuracy and can process videos in 10-15 fps.Dept. of Computer Science. Paper copy at Leddy Library: Theses & Major Papers - Basement, West Bldg. / Call Number: Thesis2006 .A42. Source: Masters Abstracts International, Volume: 45-01, page: 0347. Thesis (M.Sc.)--University of Windsor (Canada), 2006

    Vision-Language Models can Identify Distracted Driver Behavior from Naturalistic Videos

    Full text link
    Recognizing the activities, causing distraction, in real-world driving scenarios is critical for ensuring the safety and reliability of both drivers and pedestrians on the roadways. Conventional computer vision techniques are typically data-intensive and require a large volume of annotated training data to detect and classify various distracted driving behaviors, thereby limiting their efficiency and scalability. We aim to develop a generalized framework that showcases robust performance with access to limited or no annotated training data. Recently, vision-language models have offered large-scale visual-textual pretraining that can be adapted to task-specific learning like distracted driving activity recognition. Vision-language pretraining models, such as CLIP, have shown significant promise in learning natural language-guided visual representations. This paper proposes a CLIP-based driver activity recognition approach that identifies driver distraction from naturalistic driving images and videos. CLIP's vision embedding offers zero-shot transfer and task-based finetuning, which can classify distracted activities from driving video data. Our results show that this framework offers state-of-the-art performance on zero-shot transfer and video-based CLIP for predicting the driver's state on two public datasets. We propose both frame-based and video-based frameworks developed on top of the CLIP's visual representation for distracted driving detection and classification task and report the results.Comment: 15 pages, 10 figure

    A meta-model for dataflow-based rules in smart environments: Evaluating user comprehension and performance

    Full text link
    NOTICE: this is the author’s version of a work that was accepted for publication in Science of Computer Programming Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. A definitive version was subsequently published in Science of Computer Programming, [Volume 78, Issue 10, 1 October 2013, Pages 1930–1950] DOI 10.1016/j.scico.2012.06.010A considerable part of the behavior in smart environments relies on event-driven and rule specification. Rules are the mechanism most often used to enable user customization of the environment. However, the expressiveness of the rules available to users in editing and other tools is usually either limited or the available rule editing interfaces are not designed for end-users with low skills in programming. This means we have to look for interaction techniques and new ways to define user customization rules. This paper describes a generic and flexible meta-model to support expressive rules enhanced with data flow expressions that will graphically support the definition of rules without writing code. An empirical study was conducted on the ease of understanding of the visual data flow expressions, which are the key elements in our rule proposal. The visual dataflow language was compared to its corresponding textual version in terms of comprehension and ease of learning by teenagers in exercises involving calculations, modifications, writing and detecting equivalences in expressions in both languages. Although the subjects had some previous experience in editing mathematical expressions on spreadsheets, the study found their performance with visual dataflows to be significantly better in calculation and modification exercises. This makes our dataflow approach a promising mechanism for expressing user-customized reactive behavior in Ambient Intelligence (AmI) environments. The performance of the rule matching processor was validated by means of two stress tests to ensure that the meta-model approach adopted would be able to scale up with the number of types and instances in the space. © 2012 Elsevier B.V. All rights reserved.This work received financial support from the Spanish Ministry of Education under the National Strategic Program of Research and Project TSI2010-20488. Our thanks also go to the high school "Collegi Parroquial D. Jose Lluch - Alboraya", especially to the teachers and students that participated in the empirical study reported in this paper. A. Catala is supported by an FPU fellowship from the Ministry of Education of Spain with reference AP2006-00181.Catalá Bolós, A.; Pons, P.; Jaén Martínez, FJ.; Mocholi Agües, JA.; Navarro Martínez, EM. (2013). A meta-model for dataflow-based rules in smart environments: Evaluating user comprehension and performance. Science of Computer Programming. 78(10):1930-1950. doi:10.1016/j.scico.2012.06.010S19301950781

    Markerless tumor-tracking algorithm using prior 4D-CBCT

    Get PDF
    Respiratory motion management is a huge challenge in radiation therapy. Respiratory motion induces temporal anatomic changes that distort the tumor volume and its position. In this study, a markerless tumor-tracking algorithm was investigated by performing phase recognition during stereotactic body radiation therapy (SBRT) using four-dimensional cone-beam computer tomography (4D-CBCT) obtained at patient registration, and in-treatment cone-beam projection images. The data for 20 treatment sessions (five lung cancer patients) were selected for this study. Three of the patients were treated with conventional flattening filter (FF) beams, and the other two were treated with flattening filter-free (FFF) beams. Prior to treatment, 4D-CBCT was acquired to create the template projection images for 10 phases. In-treatment images were obtained at near real time during treatment. Template-based phase recognition was performed for 4D-CBCT re-projected templates using prior 4D-CBCT based phase recognition algorithm and was compared with the results generated by the Amsterdam Shroud (AS) technique. Visual verification technique was used for the verification of the phase recognition and AS technique at certain tumor-visible angles. Offline template matching analysis using the cross-correlation indicated that phase recognition performed using the prior 4D-CBCT and visual verification matched up to 97.5% in the case of FFF, and 95% in the case of FF, whereas the AS technique matched 83.5% with visual verification for FFF and 93% for FF. Markerless tumor tracking based on phase recognition using prior 4D-CBCT has been developed successfully. This is the first study that reports on the use of prior 4D-CBCT based on normalized cross-correlation technique for phase recognition

    Inviwo -- A Visualization System with Usage Abstraction Levels

    Full text link
    The complexity of today's visualization applications demands specific visualization systems tailored for the development of these applications. Frequently, such systems utilize levels of abstraction to improve the application development process, for instance by providing a data flow network editor. Unfortunately, these abstractions result in several issues, which need to be circumvented through an abstraction-centered system design. Often, a high level of abstraction hides low level details, which makes it difficult to directly access the underlying computing platform, which would be important to achieve an optimal performance. Therefore, we propose a layer structure developed for modern and sustainable visualization systems allowing developers to interact with all contained abstraction levels. We refer to this interaction capabilities as usage abstraction levels, since we target application developers with various levels of experience. We formulate the requirements for such a system, derive the desired architecture, and present how the concepts have been exemplary realized within the Inviwo visualization system. Furthermore, we address several specific challenges that arise during the realization of such a layered architecture, such as communication between different computing platforms, performance centered encapsulation, as well as layer-independent development by supporting cross layer documentation and debugging capabilities
    • …
    corecore