285 research outputs found
Plane rectification through robust vanishing point tracking using the expectation-maximization algorithm
This paper introduces a new strategy for plane rectification in sequences of images, based on the Expectation-Maximization (EM) algorithm. Our approach is able to compute simultaneously the parameters of the dominant vanishing point in the image plane and the most significant lines passing through it. It is based on a novel definition of the likelihood distribution of the gradient image considering both the position and the orientation of the gradient pixels. Besides, the mixture model in which the EM algorithm operates is extended, compared to other works, to consider an additional component to control the presence of outliers. Some synthetic data tests are described to show the robustness and efficiency of the proposed method. The plane rectification results show that the method is able to remove the perspective and affine distortion of real traffic sequences without the need to compute two vanishing point
Road environment modeling using robust perspective analysis and recursive Bayesian segmentation
Recently, vision-based advanced driver-assistance systems (ADAS) have received a new increased interest to enhance driving safety. In particular, due to its high performance–cost ratio, mono-camera systems are arising as the main focus of this field of work. In this paper we present a novel on-board road modeling and vehicle detection system, which is a part of the result of the European I-WAY project. The system relies on a robust estimation of the perspective of the scene, which adapts to the dynamics of the vehicle and generates a stabilized rectified image of the road plane. This rectified plane is used by a recursive Bayesian classi- fier, which classifies pixels as belonging to different classes corresponding to the elements of interest of the scenario. This stage works as an intermediate layer that isolates subsequent modules since it absorbs the inherent variability of the scene. The system has been tested on-road, in different scenarios, including varied illumination and adverse weather conditions, and the results have been proved to be remarkable even for such complex scenarios
Calibration and Sensitivity Analysis of a Stereo Vision-Based Driver Assistance System
Az http://intechweb.org/ alatti "Books" fül alatt kell rákeresni a "Stereo Vision" címre és az 1. fejezetre
On-road visual vehicle tracking using Markov chain Monte Carlo with metropolis sampling
In this study, a method for vehicle tracking through video analysis based on Markov chain Monte Carlo (MCMC) particle filtering with metropolis sampling is proposed. The method handles multiple targets with low computational requirements and is, therefore, ideally suited for advanced-driver assistance systems that involve real-time operation. The method exploits the removed perspective domain given by inverse perspective mapping (IPM) to define a fast and efficient likelihood model. Additionally, the method encompasses an interaction model using Markov Random Fields (MRF) that allows treatment of dependencies between the motions of targets. The proposed method is tested in highway sequences and compared to state-of-the-art methods for vehicle tracking, i.e., independent target tracking with Kalman filtering (KF) and joint tracking with particle filtering. The results showed fewer tracking failures using the proposed method
Real-Time Traffic Analysis using Deep Learning Techniques and UAV based Video
In urban environments there are daily issues of traffic congestion which city authorities need to address. Realtime analysis of traffic flow information is crucial for efficiently managing urban traffic. This paper aims to conduct traffic analysis using UAV-based videos and deep learning techniques. The road traffic video is collected by using a position-fixed UAV. The most recent deep learning methods are applied to identify the moving objects in videos. The relevant mobility metrics are calculated to conduct traffic analysis and measure the consequences of traffic congestion. The proposed approach is validated with the manual analysis results and the visualization results. The traffic analysis process is real-time in terms of the pre-trained model used
Trajectory based video analysis in multi-camera setups
PhDThis thesis presents an automated framework for activity analysis in multi-camera
setups. We start with the calibration of cameras particularly without overlapping
views. An algorithm is presented that exploits trajectory observations in each view
and works iteratively on camera pairs. First outliers are identified and removed
from observations of each camera. Next, spatio-temporal information derived from
the available trajectory is used to estimate unobserved trajectory segments in areas
uncovered by the cameras. The unobserved trajectory estimates are used to estimate
the relative position of each camera pair, whereas the exit-entrance direction of
each object is used to estimate their relative orientation. The process continues and
iteratively approximates the configuration of all cameras with respect to each other.
Finally, we refi ne the initial configuration estimates with bundle adjustment, based
on the observed and estimated trajectory segments. For cameras with overlapping
views, state-of-the-art homography based approaches are used for calibration.
Next we establish object correspondence across multiple views. Our algorithm
consists of three steps, namely association, fusion and linkage. For association,
local trajectory pairs corresponding to the same physical object are estimated using
multiple spatio-temporal features on a common ground plane. To disambiguate
spurious associations, we employ a hybrid approach that utilises the matching results
on the image plane and ground plane. The trajectory segments after association
are fused by adaptive averaging. Trajectory linkage then integrates segments and generates a single trajectory of an object across the entire observed area.
Finally, for activities analysis clustering is applied on complete trajectories. Our
clustering algorithm is based on four main steps, namely the extraction of a set of
representative trajectory features, non-parametric clustering, cluster merging and
information fusion for the identification of normal and rare object motion patterns.
First we transform the trajectories into a set of feature spaces on which Meanshift
identi es the modes and the corresponding clusters. Furthermore, a merging
procedure is devised to re fine these results by combining similar adjacent clusters.
The fi nal common patterns are estimated by fusing the clustering results across all
feature spaces. Clusters corresponding to reoccurring trajectories are considered as
normal, whereas sparse trajectories are associated to abnormal and rare events.
The performance of the proposed framework is evaluated on standard data-sets
and compared with state-of-the-art techniques. Experimental results show that
the proposed framework outperforms state-of-the-art algorithms both in terms of
accuracy and robustness
3D Modelling for Improved Visual Traffic Analytics
Advanced Traffic Management Systems utilize diverse types of sensor networks with the goal of improving mobility and safety of transportation systems. These systems require information about the state of the traffic configuration, including volume, vehicle speed, density, and incidents, which are useful in applications such as urban planning, collision avoidance systems, and emergency vehicle notification systems, to name a few.
Sensing technologies are an important part of Advanced Traffic Management Systems that enable the estimation of the traffic state. Inductive Loop Detectors are often used to sense vehicles on highway roads. Although this technology has proven to be effective, it has limitations. Their installation and replacement cost is high and causes traffic disruptions, and their sensing modality provides very limited information about the vehicles being sensed. No vehicle appearance information is available. Traffic camera networks are also used in advanced traffic monitoring centers where the cameras are controlled by a remote operator. The amount of visual information provided by such cameras can be overwhelmingly large, which may cause the operators to miss important traffic events happening in the field.
This dissertation focuses on visual traffic surveillance for Advanced Traffic Management Systems. The focus is on the research and development of computer vision algorithms that contribute to the automation of highway traffic analytics systems that require estimates of traffic volume and density.
This dissertation makes three contributions:
The first contribution is an integrated vision surveillance system called
3DTown, where cameras installed at a university campus together with algorithms are used to produce vehicle and pedestrian detections to augment a 3D model of the university with dynamic information from the scene.
A second major contribution is a technique for extracting road lines from highway images that are used to estimate the tilt angle and the focal length of the camera. This technique is useful when the operator changes the camera pose.
The third major contribution is a method to automatically extract the
active road lanes and model the vehicles in 3D to improve the vehicle
count estimation by individuating 2D segments of imaged vehicles that
have been merged due to occlusions
Unsupervised Methods for Camera Pose Estimation and People Counting in Crowded Scenes
Most visual crowd counting methods rely on training with labeled data to learn a mapping between features in the image and the number of people in the scene. However, the exact nature of this mapping may change as a function of different scene and viewing conditions, limiting the ability of such supervised systems to generalize to novel conditions, and thus preventing broad deployment. Here I propose an alternative, unsupervised strategy anchored on a 3D simulation that automatically learns how groups of people appear in the image and adapts to the signal processing parameters of the current viewing scenario. To implement this 3D strategy, knowledge of the camera parameters is required. Most methods for automatic camera calibration make assumptions about regularities in scene structure or motion patterns, which do not always apply. I propose a novel motion based approach for recovering camera tilt that does not require tracking. Having an automatic camera calibration method allows for the implementation of an accurate crowd counting algorithm that reasons in 3D. The system is evaluated on various datasets and compared against state-of-art methods
Omnidirectional Stereo Vision for Autonomous Vehicles
Environment perception with cameras is an important requirement for many applications for autonomous vehicles and robots. This work presents a stereoscopic omnidirectional camera system for autonomous vehicles which resolves the problem of a limited field of view and provides a 360° panoramic view of the environment. We present a new projection model for these cameras and show that the camera setup overcomes major drawbacks of traditional perspective cameras in many applications
Breast cancer : model reconstruction and image registration from segmented deformed image using visual and force based analysis
Breast lesion localization using tactile imaging is a new and developing direction in medical science. To achieve the goal, proper image reconstruction and image registration can be a valuable asset. In this paper, a new approach of the segmentation-based image surface reconstruction algorithm is used to reconstruct the surface of a breast phantom. In breast tissue, the sub-dermal vein network is used as a distinguishable pattern for reconstruction. The proposed image capturing device contacts the surface of the phantom, and surface deformation will occur due to applied force at the time of scanning. A novel force based surface rectification system is used to reconstruct a deformed surface image to its original structure. For the construction of the full surface from rectified images, advanced affine scale-invariant feature transform (A-SIFT) is proposed to reduce the affine effect in time when data capturing. Camera position based image stitching approach is applied to construct the final original non-rigid surface. The proposed model is validated in theoretical models and real scenarios, to demonstrate its advantages with respect to competing methods. The result of the proposed method, applied to path reconstruction, ends with a positioning accuracy of 99.7%
- …