3,236 research outputs found

    STV-based Video Feature Processing for Action Recognition

    Get PDF
    In comparison to still image-based processes, video features can provide rich and intuitive information about dynamic events occurred over a period of time, such as human actions, crowd behaviours, and other subject pattern changes. Although substantial progresses have been made in the last decade on image processing and seen its successful applications in face matching and object recognition, video-based event detection still remains one of the most difficult challenges in computer vision research due to its complex continuous or discrete input signals, arbitrary dynamic feature definitions, and the often ambiguous analytical methods. In this paper, a Spatio-Temporal Volume (STV) and region intersection (RI) based 3D shape-matching method has been proposed to facilitate the definition and recognition of human actions recorded in videos. The distinctive characteristics and the performance gain of the devised approach stemmed from a coefficient factor-boosted 3D region intersection and matching mechanism developed in this research. This paper also reported the investigation into techniques for efficient STV data filtering to reduce the amount of voxels (volumetric-pixels) that need to be processed in each operational cycle in the implemented system. The encouraging features and improvements on the operational performance registered in the experiments have been discussed at the end

    Quantification of cortical folding using MR image data

    Get PDF
    The cerebral cortex is a thin layer of tissue lining the brain where neural circuits perform important high level functions including sensory perception, motor control and language processing. In the third trimester the fetal cortex folds rapidly from a smooth sheet into a highly convoluted arrangement of gyri and sulci. Premature birth is a high risk factor for poor neurodevelopmental outcome and has been associated with abnormal cortical development, however the nature of the disruption to developmental processes is not fully understood. Recent developments in magnetic resonance imaging have allowed the acquisition of high quality brain images of preterms and also fetuses in-utero. The aim of this thesis is to develop techniques which quantify folding from these images in order to better understand cortical development in these two populations. A framework is presented that quantifies global and regional folding using curvature-based measures. This methodology was applied to fetuses over a wide gestational age range (21.7 to 38.9 weeks) for a large number of subjects (N = 80) extending our understanding of how the cortex folds through this critical developmental period. The changing relationship between the folding measures and gestational age was modelled with a Gompertz function which allowed an accurate prediction of physiological age. A spectral-based method is outlined for constructing a spatio-temporal surface atlas (a sequence of mean cortical surface meshes for weekly intervals). A key advantage of this method is the ability to do group-wise atlasing without bias to the anatomy of an initial reference subject. Mean surface templates were constructed for both fetuses and preterms allowing a preliminary comparison of mean cortical shape over the postmenstrual age range 28-36 weeks. Displacement patterns were revealed which intensified with increasing prematurity, however more work is needed to evaluate the reliability of these findings.Open Acces

    Online Spatio-Temporal Gaussian Process Experts with Application to Tactile Classification

    No full text

    Video object segmentation aggregation

    Full text link
    © 2016 IEEE. We present an approach for unsupervised object segmentation in unconstrained videos. Driven by the latest progress in this field, we argue that segmentation performance can be largely improved by aggregating the results generated by state-of-the-art algorithms. Initially, objects in individual frames are estimated through a per-frame aggregation procedure using majority voting. While this can predict relatively accurate object location, the initial estimation fails to cover the parts that are wrongly labeled by more than half of the algorithms. To address this, we build a holistic appearance model using non-local appearance cues by linear regression. Then, we integrate the appearance priors and spatio-temporal information into an energy minimization framework to refine the initial estimation. We evaluate our method on challenging benchmark videos and demonstrate that it outperforms state-of-the-art algorithms
    corecore