67 research outputs found
Frequency domain subpixel registration using HOG phase correlation
We present a novel frequency-domain image registration technique, which employs histograms of oriented gradients providing subpixel estimates. Our method involves image filtering using dense Histogram of Oriented Gradients (HOG), which provides an advanced representation of the images coping with real-world registration problems such as non-overlapping regions and small deformations. The proposed representation retains the orientation information and the corresponding weights in a multi-dimensional representation. Furthermore, due to the overlapping local contrast normalization characteristic of HOG, the proposed Histogram of Oriented Gradients - Phase Correlation (HOG-PC) method improves significantly the estimated motion parameters in small size blocks. Experiments using sequences with and without ground truth including both global and local/multiple motions demonstrate that the proposed method out- performs the state-of-the-art in frequency-domain motion estimation, in the shape of phase correlation, in terms of subpixel accuracy and motion compensation prediction for a range of test material, block sizes and motion scenarios
Phase amplified correlation for improved sub-pixel motion estimation
Phase correlation (PC) is widely employed by several sub-pixel motion estimation techniques in an attempt to accurately and robustly detect the displacement between two images. To achieve sub-pixel accuracy, these techniques employ interpolation methods and function-fitting approaches on the cross-correlation function derived from the PC core. However, such motion estimation techniques still present a lower bound of accuracy that cannot be overcome. To allow room for further improvements, we propose in this paper the enhancement of the sub-pixel accuracy of motion estimation techniques by employing a completely different approach: the concept of motion magnification. To this end, we propose the novel phase amplified correlation (PAC) that integrates motion magnification between two compared images inside the phase correlation part of frequencybased motion estimation algorithms and thus directly substitutes the PC core. The experimentation on magnetic resonance (MR) images and real video sequences demonstrates the ability of the proposed PAC core to make subtle motions highly distinguishable and improve the sub-pixel accuracy of frequency-based motion estimation techniques
Investigation of Computer Vision Concepts and Methods for Structural Health Monitoring and Identification Applications
This study presents a comprehensive investigation of methods and technologies for developing a computer vision-based framework for Structural Health Monitoring (SHM) and Structural Identification (St-Id) for civil infrastructure systems, with particular emphasis on various types of bridges. SHM is implemented on various structures over the last two decades, yet, there are some issues such as considerable cost, field implementation time and excessive labor needs for the instrumentation of sensors, cable wiring work and possible interruptions during implementation. These issues make it only viable when major investments for SHM are warranted for decision making. For other cases, there needs to be a practical and effective solution, which computer-vision based framework can be a viable alternative. Computer vision based SHM has been explored over the last decade. Unlike most of the vision-based structural identification studies and practices, which focus either on structural input (vehicle location) estimation or on structural output (structural displacement and strain responses) estimation, the proposed framework combines the vision-based structural input and the structural output from non-contact sensors to overcome the limitations given above. First, this study develops a series of computer vision-based displacement measurement methods for structural response (structural output) monitoring which can be applied to different infrastructures such as grandstands, stadiums, towers, footbridges, small/medium span concrete bridges, railway bridges, and long span bridges, and under different loading cases such as human crowd, pedestrians, wind, vehicle, etc. Structural behavior, modal properties, load carrying capacities, structural serviceability and performance are investigated using vision-based methods and validated by comparing with conventional SHM approaches. In this study, some of the most famous landmark structures such as long span bridges are utilized as case studies. This study also investigated the serviceability status of structures by using computer vision-based methods. Subsequently, issues and considerations for computer vision-based measurement in field application are discussed and recommendations are provided for better results. This study also proposes a robust vision-based method for displacement measurement using spatio-temporal context learning and Taylor approximation to overcome the difficulties of vision-based monitoring under adverse environmental factors such as fog and illumination change. In addition, it is shown that the external load distribution on structures (structural input) can be estimated by using visual tracking, and afterward load rating of a bridge can be determined by using the load distribution factors extracted from computer vision-based methods. By combining the structural input and output results, the unit influence line (UIL) of structures are extracted during daily traffic just using cameras from which the external loads can be estimated by using just cameras and extracted UIL. Finally, the condition assessment at global structural level can be achieved using the structural input and output, both obtained from computer vision approaches, would give a normalized response irrespective of the type and/or load configurations of the vehicles or human loads
Recommended from our members
Computer Vision Sensing Systems for Structural Health Monitoring in Challenging Field Conditions
Computer vision sensing techniques enable easy-to-install and remote non-contact monitoring of structures and have great potentials in field applications. This study will develop/implement novel computer vision techniques for two sensing systems for monitoring different aspects of infrastructures in challenging field conditions. The dissertation is therefore composed of two parts: robust measurement of global multi-point structural displacements, and accurate and robust monitoring of local surface displacements/strains.
Computer vision based displacement measurement has become popular in the recent decade. The first part presents InnoVision, a vision sensing system developed to address a number of challenging problems associated with applying vision sensors to the measurement of multi-point structural displacement in field conditions that are rarely comprehensively studied in the literature. The challenging problems include tracking low-contrast natural targets on the structural surface, insufficient resolution for long distance measurement, inevitable camera vibration, and image distortion due to heat haze in hot weather. Several techniques are developed in InnoVision to tackle these challenges. Laboratory and field tests are conducted to evaluate the performance of these techniques.
In the second part, another vision sensing system SurfaceVision is developed for accurate and robust monitoring two-dimensional (2D) structural surface displacements/strains. Important structures, such as nuclear power plants, need the continuous inspection of surface conditions. As an alternative to the human inspection, conventional digital-image-correlation (DIC) based methods have been applied to surfaces painted with speckle patterns in a controlled environment. However, it is highly challenging for DIC methods to accurately measure displacement on natural concrete surfaces in outdoor conditions with changing illumination and weather conditions. Additionally, common surface displacement measurement is based on segmenting the surface image into small subsets and tracking each subset individually through template matching, the surface displacement thus obtained has obvious discontinuity and low spatial resolution. Therefore, for applicability in the outdoor environment, SurfaceVision is proposed for accurate and robust monitoring of surface displacements/strains. Advanced computer vision techniques are developed/implemented to enable surface displacement measurement with high continuity, spatial resolution, accuracy, and robustness. An intuitive strain calculation method is also developed for converting surface displacements into surface strains. A numerical simulation is formulated based on four-point bending tests to validate the accuracy and robustness of SurfaceVision in surface displacements. Four-point bending experiments using reinforced concrete specimens are conducted to demonstrate the performance of SurfaceVision under different cases of optical noises and its effectiveness in predicting crack formations
Recommended from our members
An automated image processing system for the detection of photoreceptor cells in adaptive optics retinal images
The rapid progress in Adaptive Optics (AO) imaging, in the last decades, has had a transformative impact on the entire approach underpinning the investigations of retinal tissues. Capable of imaging the retina in vivo at the cellular level, AO systems have revealed new insights into retinal structures, function, and the origins of various retinal pathologies. This has expanded the field of clinical research and opened a wide range of applications for AO imaging. The advances in image processing techniques contribute to a better observation of retinal microstructures and therefore more accurate detection of pathological conditions. The development of automated tools for processing images obtained with AO allows for objective examination of a larger number of images with time and cost savings and thus facilitates the use of AO imaging as a practical and efficient tool, by making it widely accessible to the clinical ophthalmic community.
In this work, an image processing framework is developed that allows for enhancement of AO high-resolution retinal images and accurate detection of photoreceptor cells. The proposed framework consists of several stages: image quality assessment, illumination compensation, noise suppression, image registration, image restoration, enhancement and detection of photoreceptor cells. The visibility of retinal features is improved by tackling specific components of the AO imaging system, affecting the quality of acquired retinal data. Therefore, we attempt to fully recover AO retinal images, free from any induced degradation effects. A comparative study of different methods and evaluation of their efficiency on retinal datasets is performed by assessing image quality. In order to verify the achieved results, the cone packing density distribution was calculated and correlated with statistical histological data. From the performed experiments, it can be concluded that the proposed image processing framework can effectively improve photoreceptor cell image quality and thus can serve as a platform for further investigation of retinal tissues. Quantitative analysis of the retinal images obtained with the proposed image processing framework can be used for comparison with data related to pathological retinas, as well as for understanding the effect of age and retinal pathology on cone packing density and other microstructures
Vision-based localization methods under GPS-denied conditions
This paper reviews vision-based localization methods in GPS-denied
environments and classifies the mainstream methods into Relative Vision
Localization (RVL) and Absolute Vision Localization (AVL). For RVL, we discuss
the broad application of optical flow in feature extraction-based Visual
Odometry (VO) solutions and introduce advanced optical flow estimation methods.
For AVL, we review recent advances in Visual Simultaneous Localization and
Mapping (VSLAM) techniques, from optimization-based methods to Extended Kalman
Filter (EKF) based methods. We also introduce the application of offline map
registration and lane vision detection schemes to achieve Absolute Visual
Localization. This paper compares the performance and applications of
mainstream methods for visual localization and provides suggestions for future
studies.Comment: 32 pages, 15 figure
Carried baggage detection and recognition in video surveillance with foreground segmentation
Security cameras installed in public spaces or in private organizations continuously
record video data with the aim of detecting and preventing crime. For that reason,
video content analysis applications, either for real time (i.e. analytic) or post-event
(i.e. forensic) analysis, have gained high interest in recent years. In this thesis,
the primary focus is on two key aspects of video analysis, reliable moving object
segmentation and carried object detection & identification.
A novel moving object segmentation scheme by background subtraction is presented
in this thesis. The scheme relies on background modelling which is based
on multi-directional gradient and phase congruency. As a post processing step,
the detected foreground contours are refined by classifying the edge segments as
either belonging to the foreground or background. Further contour completion
technique by anisotropic diffusion is first introduced in this area. The proposed
method targets cast shadow removal, gradual illumination change invariance, and
closed contour extraction.
A state of the art carried object detection method is employed as a benchmark
algorithm. This method includes silhouette analysis by comparing human temporal
templates with unencumbered human models. The implementation aspects of
the algorithm are improved by automatically estimating the viewing direction of
the pedestrian and are extended by a carried luggage identification module. As
the temporal template is a frequency template and the information that it provides
is not sufficient, a colour temporal template is introduced. The standard
steps followed by the state of the art algorithm are approached from a different
extended (by colour information) perspective, resulting in more accurate carried
object segmentation.
The experiments conducted in this research show that the proposed closed
foreground segmentation technique attains all the aforementioned goals. The incremental
improvements applied to the state of the art carried object detection
algorithm revealed the full potential of the scheme. The experiments demonstrate
the ability of the proposed carried object detection algorithm to supersede the
state of the art method
- …