2,003 research outputs found

    Moving-edge detection via heat flow analogy

    No full text
    In this paper, a new and automatic moving-edge detection algorithm is proposed, based on using the heat flow analogy. This algorithm starts with anisotropic heat diffusion in the spatial domain, to remove noise and sharpen region boundaries for the purpose of obtaining high quality edge data. Then, isotropic and linear heat diffusion is applied in the temporal domain to calculate the total amount of heat flow. The moving-edges are represented as the total amount of heat flow out from the reference frame. The overall process is completed by non-maxima suppression and hysteresis thresholding to obtain binary moving edges. Evaluation, on a variety of data, indicates that this approach can handle noise in the temporal domain because of the averaging inherent of isotropic heat flow. Results also show that this technique can detect moving-edges in image sequences, without background image subtraction

    Automatic Image Based Time Varying 3D Feature Extraction and Tracking

    Get PDF
    3D time-varying data sets are complex. The intrinsics of those data cannot be readily comprehended by users solely based on visual investigation. Computational tools such as feature extraction and tracking are often necessary. Until now, most existing algorithms in this domain work effectively in the object space, relying on prior knowledge of the data. How to find a more flexible and efficient method which can perform automatically to implement extraction and tracking remains an attractive topic. This thesis presents a new image-based method that extracts and tracks the 3D time- varying volume data sets. The innovation of the proposed approach is two-fold. First, all analyses are performed in the image space on volume rendered images without accessing the actual volume data itself. The image-based processing will help to both save storage space in the memory and reduce computation burden. Secondly, the new approach does not require any prior knowledge of the user-defined “feature” or a built model. All the parameters used by the algorithms are automatically determined by the system itself, thus flexibility and efficiency can be achieved at the same time. The proposed image-based feature extraction and tracking system consists of four components: feature segmentation (or extraction), feature description (or shape analysis), classification, and feature tracking. Feature segmentation is to identify and label individual features from the image so that we can describe and track them separately. We combine both region-based and edge-based segmentation approaches to implement the extraction process. Feature description is to analyze each feature and derive a vector to describe the feature such that the subsequent tracking step does not have to rely on the entire feature extracted, but instead a much smaller and informative feature descriptor. Classification is to identify the corresponding features from two consecutive image frames along both the time and the spatial domain. Feature tracking is to study and model the evolution of features based on the correspondence computation result from classification stage. Experimental results show that the image-based feature extraction and tracking system provides high fidelity with great efficiency

    Segmentation of neuroanatomy in magnetic resonance images

    Get PDF
    Segmentation in neurological Magnetic Resonance Imaging (MRI) is necessary for volume measurement, feature extraction and for the three-dimensional display of neuroanatomy. This thesis proposes several automated and semi-automated methods which offer considerable advantages over manual methods because of their lack of subjectivity, their data reduction capabilities, and the time savings they give. Work has concentrated on the use of dual echo multi-slice spin-echo data sets in order to take advantage of the intrinsically multi-parametric nature of MRI. Such data is widely acquired clinically and segmentation therefore does not require additional scans. The literature has been reviewed. Factors affecting image non-uniformity for a modem 1.5 Tesla imager have been investigated. These investigations demonstrate that a robust, fast, automatic three-dimensional non-uniformity correction may be applied to data as a pre-processing step. The merit of using an anisotropic smoothing method for noisy data has been demonstrated. Several approaches to neurological MRI segmentation have been developed. Edge-based processing is used to identify the skin (the major outer contour) and the eyes. Edge-focusing, two threshold based techniques and a fast radial CSF identification approach are proposed to identify the intracranial region contour in each slice of the data set. Once isolated, the intracranial region is further processed to identify CSF, and, depending upon the MRI pulse sequence used, the brain itself may be sub-divided into grey matter and white matter using semiautomatic contrast enhancement and clustering methods. The segmentation of Multiple Sclerosis (MS) plaques has also been considered. The utility of the stack, a data driven multi-resolution approach to segmentation, has been investigated, and several improvements to the method suggested. The factors affecting the intrinsic accuracy of neurological volume measurement in MRI have been studied and their magnitudes determined for spin-echo imaging. Geometric distortion - both object dependent and object independent - has been considered, as well as slice warp, slice profile, slice position and the partial volume effect. Finally, the accuracy of the approaches to segmentation developed in this thesis have been evaluated. Intracranial volume measurements are within 5% of expert observers' measurements, white matter volumes within 10%, and CSF volumes consistently lower than the expert observers' measurements due to the observers' inability to take the partial volume effect into account

    Geometrical-based approach for robust human image detection

    Get PDF
    In recent years, object detection and classification has been gaining more attention, thus, there are several human object detection algorithms being used to locate and recognize human objects in images. The research of image processing and analyzing based on human shape is one of the hot topic due to the wide applicability in real applications. In this paper, we present a new object classification approach. The new approach will use a simple and robust geometrical model to classify the detected object as human or non-human in the images. In the proposed approach, the object is detected. Then the detected object under different conditions can be accurately classified (i.e. human, non-human) by combining the features that are extracted from the upper portion of the contour and the proposed geometrical model parameters. A software-based simulation using Matlab was performed using INRIA dataset and the obtained results are validated by comparing with five state-of-art approaches in literature and some of the machine learning approaches such as artificial neural networks (ANN), support vector machine (SVM), and random forest (RF). The experimental results show that the proposed object classification approach is efficient and achieved a comparable accuracy to other machine learning approaches and other state-of-art approaches. Keywords: Human classification, Geometrical model, INRIA, Machine learning, SVM, ANN, Random forest

    Audio-coupled video content understanding of unconstrained video sequences

    Get PDF
    Unconstrained video understanding is a difficult task. The main aim of this thesis is to recognise the nature of objects, activities and environment in a given video clip using both audio and video information. Traditionally, audio and video information has not been applied together for solving such complex task, and for the first time we propose, develop, implement and test a new framework of multi-modal (audio and video) data analysis for context understanding and labelling of unconstrained videos. The framework relies on feature selection techniques and introduces a novel algorithm (PCFS) that is faster than the well-established SFFS algorithm. We use the framework for studying the benefits of combining audio and video information in a number of different problems. We begin by developing two independent content recognition modules. The first one is based on image sequence analysis alone, and uses a range of colour, shape, texture and statistical features from image regions with a trained classifier to recognise the identity of objects, activities and environment present. The second module uses audio information only, and recognises activities and environment. Both of these approaches are preceded by detailed pre-processing to ensure that correct video segments containing both audio and video content are present, and that the developed system can be made robust to changes in camera movement, illumination, random object behaviour etc. For both audio and video analysis, we use a hierarchical approach of multi-stage classification such that difficult classification tasks can be decomposed into simpler and smaller tasks. When combining both modalities, we compare fusion techniques at different levels of integration and propose a novel algorithm that combines advantages of both feature and decision-level fusion. The analysis is evaluated on a large amount of test data comprising unconstrained videos collected for this work. We finally, propose a decision correction algorithm which shows that further steps towards combining multi-modal classification information effectively with semantic knowledge generates the best possible results

    Basic Science to Clinical Research: Segmentation of Ultrasound and Modelling in Clinical Informatics

    Get PDF
    The world of basic science is a world of minutia; it boils down to improving even a fraction of a percent over the baseline standard. It is a domain of peer reviewed fractions of seconds and the world of squeezing every last ounce of efficiency from a processor, a storage medium, or an algorithm. The field of health data is based on extracting knowledge from segments of data that may improve some clinical process or practice guideline to improve the time and quality of care. Clinical informatics and knowledge translation provide this information in order to reveal insights to the world of improving patient treatments, regimens, and overall outcomes. In my world of minutia, or basic science, the movement of blood served an integral role. The novel detection of sound reverberations map out the landscape for my research. I have applied my algorithms to the various anatomical structures of the heart and artery system. This serves as a basis for segmentation, active contouring, and shape priors. The algorithms presented, leverage novel applications in segmentation by using anatomical features of the heart for shape priors and the integration of optical flow models to improve tracking. The presented techniques show improvements over traditional methods in the estimation of left ventricular size and function, along with plaque estimation in the carotid artery. In my clinical world of data understanding, I have endeavoured to decipher trends in Alzheimer’s disease, Sepsis of hospital patients, and the burden of Melanoma using mathematical modelling methods. The use of decision trees, Markov models, and various clustering techniques provide insights into data sets that are otherwise hidden. Finally, I demonstrate how efficient data capture from providers can achieve rapid results and actionable information on patient medical records. This culminated in generating studies on the burden of illness and their associated costs. A selection of published works from my research in the world of basic sciences to clinical informatics has been included in this thesis to detail my transition. This is my journey from one contented realm to a turbulent one

    Active modelling of virtual humans

    Get PDF
    This thesis provides a complete framework that enables the creation of photorealistic 3D human models in real-world environments. The approach allows a non-expert user to use any digital capture device to obtain four images of an individual and create a personalised 3D model, for multimedia applications. To achieve this, it is necessary that the system is automatic and that the reconstruction process is flexible to account for information that is not available or incorrectly captured. In this approach the individual is automatically extracted from the environment using constrained active B-spline templates that are scaled and automatically initialised using only image information. These templates incorporate the energy minimising framework for Active Contour Models, providing a suitable and flexible method to deal with the adjustments in pose an individual can adopt. The final states of the templates describe the individual’s shape. The contours in each view are combined to form a 3D B-spline surface that characterises an individual’s maximal silhouette equivalent. The surface provides a mould that contains sufficient information to allow for the active deformation of an underlying generic human model. This modelling approach is performed using a novel technique that evolves active-meshes to 3D for deforming the underlying human model, while adaptively constraining it to preserve its existing structure. The active-mesh approach incorporates internal constraints that maintain the structural relationship of the vertices of the human model, while external forces deform the model congruous to the 3D surface mould. The strength of the internal constraints can be reduced to allow the model to adopt the exact shape of the bounding volume or strengthened to preserve the internal structure, particularly in areas of high detail. This novel implementation provides a uniform framework that can be simply and automatically applied to the entire human model
    • 

    corecore