2,209 research outputs found

    Video Processing Analysis For Non-Invasive Fatigue Detection And Quantification

    Get PDF
    Fatigue is a common symptom of weakness either physically or mentally. These symptoms may led to a drop in motivation, weakened sensitivity, slowing of responsiveness and inability to give full attention. All of these problems can cause adverse effects, such as accidents, especially those that require full attention as drivers of vehicles, and rail operators, the pilot of an aircraft or ship operators. This research investigates systems to detect and quantify the signs of fatigue using non-invasive facial analytics. There are four main algorithms that represent the major contribution from the PhD research. These algorithms encompass facial fatigue detection and quantification system as a whole. Firstly, a new technique to detect the face is introduced. This face detection algorithm is an affiliation of colour skin segmentation technique, connected component of binary image usage, and learning machine algorithm. The introduced face detection algorithm is able to reduce the false positive detection rate by a very significant margin. For the facial fatigue detection and quantification, the major fatigue signs features are from the eye activity. A new algorithm called the , Interdependence and Adaptive Scale Mean Shift (IASMS) is presented. The IASMS is able to quantify the state of eye as well as to track non-rigid eye movement. IASMS integrates the mean shift tracking algorithm with an adaptive scale scheme, which is used to track the iris and quantify the iris size. The IASMS is associated with face detection algorithm, image enhanced scheme, eye open detection technique and iris detection method in the initialisation process. This proposed method is able to quantify the eye activities that represent the blink rate and the duration of eye closure. The third contribution is yawning analysis algorithm. Commonly yawning is detected based on a wide mouth opening. Frequently however this approach is thwarted by the common human reaction to hand-cover the mouth during yawning. In this research, a new approach to analyse yawning which takes into account the covered mouth is introduced. This algorithm combines with a new technique of mouth opening measurements, covered mouth detection, and facial distortion (wrinkles) detection. By using this proposed method, yawning is still able to detect even though the mouth is covered. In order to have reliable results from the testing and evaluating of the developed fatigue detection algorithm, the real signs of fatigue are required. This research develops a recorded face activities database of the people that experience fatigue. This fatigue database is called as the Strathclyde Fatigue Facial (SFF). To induce the fatigue signs, ethically approved sleep deprivation experiments were carried out. In these experiments twenty participants, and four sessions were undertaken, which the participant has to deprive their sleep in 0, 3, 5, and 8 hours. The participants were subsequently requested to carry out 5 cognitive tasks that are related to the sleep loss. The last contribution of this research is a technique to recognise the fatigue signs. The existing fatigue detection system is based on single classification. However, this work presents a new approach for fatigue recognition which the fatigue is classified into levels. The levels of fatigue are justified based on the sleep deprivation stages where the SFF database is fully used for training, testing and evaluation of the developed fatigue recognition algorithm. This fatigue recognition algorithm is then integrated into a Fatigue Monitoring Tool (FMT) platform. This FMT has been used to test the participant that carried out the tasks as ship crew in shipping bridge simulator

    Generalized Completed Local Binary Patterns for Time-Efficient Steel Surface Defect Classification

    Get PDF
    © 2018 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted ncomponent of this work in other works.Efficient defect classification is one of the most important preconditions to achieve online quality inspection for hot-rolled strip steels. It is extremely challenging owing to various defect appearances, large intraclass variation, ambiguous interclass distance, and unstable gray values. In this paper, a generalized completed local binary patterns (GCLBP) framework is proposed. Two variants of improved completed local binary patterns (ICLBP) and improved completed noise-invariant local-structure patterns (ICNLP) under the GCLBP framework are developed for steel surface defect classification. Different from conventional local binary patterns variants, descriptive information hidden in nonuniform patterns is innovatively excavated for the better defect representation. This paper focuses on the following aspects. First, a lightweight searching algorithm is established for exploiting the dominant nonuniform patterns (DNUPs). Second, a hybrid pattern code mapping mechanism is proposed to encode all the uniform patterns and DNUPs. Third, feature extraction is carried out under the GCLBP framework. Finally, histogram matching is efficiently accomplished by simple nearest-neighbor classifier. The classification accuracy and time efficiency are verified on a widely recognized texture database (Outex) and a real-world steel surface defect database [Northeastern University (NEU)]. The experimental results promise that the proposed method can be widely applied in online automatic optical inspection instruments for hot-rolled strip steel.Peer reviewe

    Surface Defect Classification for Hot-Rolled Steel Strips by Selectively Dominant Local Binary Patterns

    Get PDF
    Developments in defect descriptors and computer vision-based algorithms for automatic optical inspection (AOI) allows for further development in image-based measurements. Defect classification is a vital part of an optical-imaging-based surface quality measuring instrument. The high-speed production rhythm of hot continuous rolling requires an ultra-rapid response to every component as well as algorithms in AOI instrument. In this paper, a simple, fast, yet robust texture descriptor, namely selectively dominant local binary patterns (SDLBPs), is proposed for defect classification. First, an intelligent searching algorithm with a quantitative thresholding mechanism is built to excavate the dominant non-uniform patterns (DNUPs). Second, two convertible schemes of pattern code mapping are developed for binary encoding of all uniform patterns and DNUPs. Third, feature extraction is carried out under SDLBP framework. Finally, an adaptive region weighting method is built for further strengthening the original nearest neighbor classifier in the feature matching stage. The extensive experiments carried out on an open texture database (Outex) and an actual surface defect database (Dragon) indicates that our proposed SDLBP yields promising performance on both classification accuracy and time efficiencyPeer reviewe

    Adaptive visual sampling

    Get PDF
    PhDVarious visual tasks may be analysed in the context of sampling from the visual field. In visual psychophysics, human visual sampling strategies have often been shown at a high-level to be driven by various information and resource related factors such as the limited capacity of the human cognitive system, the quality of information gathered, its relevance in context and the associated efficiency of recovering it. At a lower-level, we interpret many computer vision tasks to be rooted in similar notions of contextually-relevant, dynamic sampling strategies which are geared towards the filtering of pixel samples to perform reliable object association. In the context of object tracking, the reliability of such endeavours is fundamentally rooted in the continuing relevance of object models used for such filtering, a requirement complicated by realworld conditions such as dynamic lighting that inconveniently and frequently cause their rapid obsolescence. In the context of recognition, performance can be hindered by the lack of learned context-dependent strategies that satisfactorily filter out samples that are irrelevant or blunt the potency of models used for discrimination. In this thesis we interpret the problems of visual tracking and recognition in terms of dynamic spatial and featural sampling strategies and, in this vein, present three frameworks that build on previous methods to provide a more flexible and effective approach. Firstly, we propose an adaptive spatial sampling strategy framework to maintain statistical object models for real-time robust tracking under changing lighting conditions. We employ colour features in experiments to demonstrate its effectiveness. The framework consists of five parts: (a) Gaussian mixture models for semi-parametric modelling of the colour distributions of multicolour objects; (b) a constructive algorithm that uses cross-validation for automatically determining the number of components for a Gaussian mixture given a sample set of object colours; (c) a sampling strategy for performing fast tracking using colour models; (d) a Bayesian formulation enabling models of object and the environment to be employed together in filtering samples by discrimination; and (e) a selectively-adaptive mechanism to enable colour models to cope with changing conditions and permit more robust tracking. Secondly, we extend the concept to an adaptive spatial and featural sampling strategy to deal with very difficult conditions such as small target objects in cluttered environments undergoing severe lighting fluctuations and extreme occlusions. This builds on previous work on dynamic feature selection during tracking by reducing redundancy in features selected at each stage as well as more naturally balancing short-term and long-term evidence, the latter to facilitate model rigidity under sharp, temporary changes such as occlusion whilst permitting model flexibility under slower, long-term changes such as varying lighting conditions. This framework consists of two parts: (a) Attribute-based Feature Ranking (AFR) which combines two attribute measures; discriminability and independence to other features; and (b) Multiple Selectively-adaptive Feature Models (MSFM) which involves maintaining a dynamic feature reference of target object appearance. We call this framework Adaptive Multi-feature Association (AMA). Finally, we present an adaptive spatial and featural sampling strategy that extends established Local Binary Pattern (LBP) methods and overcomes many severe limitations of the traditional approach such as limited spatial support, restricted sample sets and ad hoc joint and disjoint statistical distributions that may fail to capture important structure. Our framework enables more compact, descriptive LBP type models to be constructed which may be employed in conjunction with many existing LBP techniques to improve their performance without modification. The framework consists of two parts: (a) a new LBP-type model known as Multiscale Selected Local Binary Features (MSLBF); and (b) a novel binary feature selection algorithm called Binary Histogram Intersection Minimisation (BHIM) which is shown to be more powerful than established methods used for binary feature selection such as Conditional Mutual Information Maximisation (CMIM) and AdaBoost

    Comprehensive Survey and Analysis of Techniques, Advancements, and Challenges in Video-Based Traffic Surveillance Systems

    Get PDF
    The challenges inherent in video surveillance are compounded by a several factors, like dynamic lighting conditions, the coordination of object matching, diverse environmental scenarios, the tracking of heterogeneous objects, and coping with fluctuations in object poses, occlusions, and motion blur. This research endeavor aims to undertake a rigorous and in-depth analysis of deep learning- oriented models utilized for object identification and tracking. Emphasizing the development of effective model design methodologies, this study intends to furnish a exhaustive and in-depth analysis of object tracking and identification models within the specific domain of video surveillance

    A framework for cardio-pulmonary resuscitation (CPR) scene retrieval from medical simulation videos based on object and activity detection.

    Get PDF
    In this thesis, we propose a framework to detect and retrieve CPR activity scenes from medical simulation videos. Medical simulation is a modern training method for medical students, where an emergency patient condition is simulated on human-like mannequins and the students act upon. These simulation sessions are recorded by the physician, for later debriefing. With the increasing number of simulation videos, automatic detection and retrieval of specific scenes became necessary. The proposed framework for CPR scene retrieval, would eliminate the conventional approach of using shot detection and frame segmentation techniques. Firstly, our work explores the application of Histogram of Oriented Gradients in three dimensions (HOG3D) to retrieve the scenes containing CPR activity. Secondly, we investigate the use of Local Binary Patterns in Three Orthogonal Planes (LBPTOP), which is the three dimensional extension of the popular Local Binary Patterns. This technique is a robust feature that can detect specific activities from scenes containing multiple actors and activities. Thirdly, we propose an improvement to the above mentioned methods by a combination of HOG3D and LBP-TOP. We use decision level fusion techniques to combine the features. We prove experimentally that the proposed techniques and their combination out-perform the existing system for CPR scene retrieval. Finally, we devise a method to detect and retrieve the scenes containing the breathing bag activity, from the medical simulation videos. The proposed framework is tested and validated using eight medical simulation videos and the results are presented
    corecore