123,924 research outputs found
Investigation into Self-calibration Methods for the Vexcel UltraCam D Digital Aerial Camera
This paper provides an investigation into the camera calibration of a Vexcel UltraCam D digital aerial camera which was undertaken as part of the EuroSDR Digital Camera Calibration project. This paper will present results from two flights flown over a test site at Fredrikstad-Norway using established camera calibration techniques. Furthermore, it proposes an alternative approach. The "new" multi cone digital camera systems are geometrically complex. The image used for photogrammetric analysis is made up of a number of images produced by a cluster of camera cones and possibly various groups of CCD arrays. This produces a resultant image which is not just based on traditional single lens/focal plane camera geometries, but depends on the joining of images from multiple lens (different perspectives), handling groups of focal planes and the matching of overlapping image areas. Some of the requirements from camera calibration such as stability can only be determined through long-term experience/research and some can be determined through investigation and short-term research such as the calibration parameters. The methodology used in this research for assessing the camera calibration is based on self-calibration using the Collinearity Equations. The analysis was undertaken in order to try to identify any systematic patterns in the resulting image residuals. By identifying and quantifying the systematic residuals, a new calibration method is proposed that recomputes the bundle adjustment based on the analysis of the systematic residual patterns. Only very small systematic patterns could be visually identified in small areas of the images. The existing self-calibration methods and the new approach have made a small improvement on the results. The new calibration approach for the low flight has been particularly beneficial in improving the RMSE in Z and reducing image residuals. However, the method was less successful at improving the high flown results. This approach has shown that it has potential but needs further investigation to fully assess its capabilities
How to turn your camera into a perfect pinhole model
Camera calibration is a first and fundamental step in various computer vision
applications. Despite being an active field of research, Zhang's method remains
widely used for camera calibration due to its implementation in popular
toolboxes. However, this method initially assumes a pinhole model with
oversimplified distortion models. In this work, we propose a novel approach
that involves a pre-processing step to remove distortions from images by means
of Gaussian processes. Our method does not need to assume any distortion model
and can be applied to severely warped images, even in the case of multiple
distortion sources, e.g., a fisheye image of a curved mirror reflection. The
Gaussian processes capture all distortions and camera imperfections, resulting
in virtual images as though taken by an ideal pinhole camera with square
pixels. Furthermore, this ideal GP-camera only needs one image of a square grid
calibration pattern. This model allows for a serious upgrade of many algorithms
and applications that are designed in a pure projective geometry setting but
with a performance that is very sensitive to nonlinear lens distortions. We
demonstrate the effectiveness of our method by simplifying Zhang's calibration
method, reducing the number of parameters and getting rid of the distortion
parameters and iterative optimization. We validate by means of synthetic data
and real world images. The contributions of this work include the construction
of a virtual ideal pinhole camera using Gaussian processes, a simplified
calibration method and lens distortion removal.Comment: 15 pages, 3 figures, conference CIAR
Monitoring Activities from Multiple Video Streams: Establishing a Common Coordinate Frame
Passive monitoring of large sites typically requires coordination between multiple cameras, which in turn requires methods for automatically relating events between distributed cameras. This paper tackles the problem of self-calibration of multiple cameras which are very far apart, using feature correspondences to determine the camera geometry. The key problem is finding such correspondences. Since the camera geometry and photometric characteristics vary considerably between images, one cannot use brightness and/or proximity constraints. Instead we apply planar geometric constraints to moving objects in the scene in order to align the scene"s ground plane across multiple views. We do not assume synchronized cameras, and we show that enforcing geometric constraints enables us to align the tracking data in time. Once we have recovered the homography which aligns the planar structure in the scene, we can compute from the homography matrix the 3D position of the plane and the relative camera positions. This in turn enables us to recover a homography matrix which maps the images to an overhead view. We demonstrate this technique in two settings: a controlled lab setting where we test the effects of errors in internal camera calibration, and an uncontrolled, outdoor setting in which the full procedure is applied to external camera calibration and ground plane recovery. In spite of noise in the internal camera parameters and image data, the system successfully recovers both planar structure and relative camera positions in both settings
Automatic multi-camera extrinsic parameter calibration based on pedestrian torsors
Extrinsic camera calibration is essential for any computer vision task in a camera network. Typically, researchers place a calibration object in the scene to calibrate all the cameras in a camera network. However, when installing cameras in the field, this approach can be costly and impractical, especially when recalibration is needed. This paper proposes a novel, accurate and fully automatic extrinsic calibration framework for camera networks with partially overlapping views. The proposed method considers the pedestrians in the observed scene as the calibration objects and analyzes the pedestrian tracks to obtain extrinsic parameters. Compared to the state of the art, the new method is fully automatic and robust in various environments. Our method detect human poses in the camera images and then models walking persons as vertical sticks. We apply a brute-force method to determines the correspondence between persons in multiple camera images. This information along with 3D estimated locations of the top and the bottom of the pedestrians are then used to compute the extrinsic calibration matrices. We also propose a novel method to calibrate the camera network by only using the top and centerline of the person when the bottom of the person is not available in heavily occluded scenes. We verified the robustness of the method in different camera setups and for both single and multiple walking people. The results show that the triangulation error of a few centimeters can be obtained. Typically, it requires less than one minute of observing the walking people to reach this accuracy in controlled environments. It also just takes a few minutes to collect enough data for the calibration in uncontrolled environments. Our proposed method can perform well in various situations such as multi-person, occlusions, or even at real intersections on the street
Online Camera-to-ground Calibration for Autonomous Driving
Online camera-to-ground calibration is to generate a non-rigid body
transformation between the camera and the road surface in a real-time manner.
Existing solutions utilize static calibration, suffering from environmental
variations such as tire pressure changes, vehicle loading volume variations,
and road surface diversity. Other online solutions exploit the usage of road
elements or photometric consistency between overlapping views across images,
which require continuous detection of specific targets on the road or
assistance with multiple cameras to facilitate calibration. In our work, we
propose an online monocular camera-to-ground calibration solution that does not
utilize any specific targets while driving. We perform a coarse-to-fine
approach for ground feature extraction through wheel odometry and estimate the
camera-to-ground calibration parameters through a sliding-window-based factor
graph optimization. Considering the non-rigid transformation of
camera-to-ground while driving, we provide metrics to quantify calibration
performance and stopping criteria to report/broadcast our satisfying
calibration results. Extensive experiments using real-world data demonstrate
that our algorithm is effective and outperforms state-of-the-art techniques
Use of stereo camera systems for assessment of rockfish abundance in untrawlable areas and for recording pollock behavior during midwater trawls
We describe the application of two types of stereo camera
systems in fisheries research, including the design, calibration, analysis techniques, and precision of the data
obtained with these systems. The first is a stereo video system deployed by using a quick-responding winch with a
live feed to provide species- and size- composition data adequate to produce acoustically based biomass estimates
of rockfish. This system was tested on the eastern Bering Sea slope where rockfish were measured. Rockfish sizes were similar to those sampled with a bottom trawl and the relative error in multiple measurements of the same rockfish in multiple still-frame images was small. Measurement errors of up to 5.5% were found on a calibration target of known size. The second system consisted of a pair of still-image digital cameras mounted
inside a midwater trawl. Processing of the stereo images allowed fish length, fish orientation in relation to the camera platform, and relative distance of the fish to the trawl netting to be determined. The video system was useful for surveying fish in Alaska, but it could also be used
broadly in other situations where it is difficult to obtain species-composition or size-composition information.
Likewise, the still-image system could be used for fisheries research to obtain data on size, position, and
orientation of fish
Lens Distortion Calibration Using Point Correspondences
This paper describes a new method for lens distortion calibration using only point correspondences in multiple views, without the need to know either the 3D location of the points or the camera locations. The standard lens distortion model is a model of the deviations of a real camera from the ideal pinhole or projective camera model.Given multiple views of a set of corresponding points taken by ideal pinhole cameras there exist epipolar and trilinear constraints among pairs and triplets of these views. In practice, due to noise in the feature detection and due to lens distortion these constraints do not hold exactly and we get some error. The calibration is a search for the lens distortion parameters that minimize this error. Using simulation and experimental results with real images we explore the properties of this method. We describe the use of this method with the standard lens distortion model, radial and decentering, but it could also be used with any other parametric distortion models. Finally we demonstrate that lens distortion calibration improves the accuracy of 3D reconstruction
- …