1,483 research outputs found

    Tracking Target Signal Strengths on a Grid using Sparsity

    Get PDF
    Multi-target tracking is mainly challenged by the nonlinearity present in the measurement equation, and the difficulty in fast and accurate data association. To overcome these challenges, the present paper introduces a grid-based model in which the state captures target signal strengths on a known spatial grid (TSSG). This model leads to \emph{linear} state and measurement equations, which bypass data association and can afford state estimation via sparsity-aware Kalman filtering (KF). Leveraging the grid-induced sparsity of the novel model, two types of sparsity-cognizant TSSG-KF trackers are developed: one effects sparsity through â„“1\ell_1-norm regularization, and the other invokes sparsity as an extra measurement. Iterative extended KF and Gauss-Newton algorithms are developed for reduced-complexity tracking, along with accurate error covariance updates for assessing performance of the resultant sparsity-aware state estimators. Based on TSSG state estimates, more informative target position and track estimates can be obtained in a follow-up step, ensuring that track association and position estimation errors do not propagate back into TSSG state estimates. The novel TSSG trackers do not require knowing the number of targets or their signal strengths, and exhibit considerably lower complexity than the benchmark hidden Markov model filter, especially for a large number of targets. Numerical simulations demonstrate that sparsity-cognizant trackers enjoy improved root mean-square error performance at reduced complexity when compared to their sparsity-agnostic counterparts.Comment: Submitted to IEEE Trans. on Signal Processin

    Camera Marker Networks for Pose Estimation and Scene Understanding in Construction Automation and Robotics.

    Full text link
    The construction industry faces challenges that include high workplace injuries and fatalities, stagnant productivity, and skill shortage. Automation and Robotics in Construction (ARC) has been proposed in the literature as a potential solution that makes machinery easier to collaborate with, facilitates better decision-making, or enables autonomous behavior. However, there are two primary technical challenges in ARC: 1) unstructured and featureless environments; and 2) differences between the as-designed and the as-built. It is therefore impossible to directly replicate conventional automation methods adopted in industries such as manufacturing on construction sites. In particular, two fundamental problems, pose estimation and scene understanding, must be addressed to realize the full potential of ARC. This dissertation proposes a pose estimation and scene understanding framework that addresses the identified research gaps by exploiting cameras, markers, and planar structures to mitigate the identified technical challenges. A fast plane extraction algorithm is developed for efficient modeling and understanding of built environments. A marker registration algorithm is designed for robust, accurate, cost-efficient, and rapidly reconfigurable pose estimation in unstructured and featureless environments. Camera marker networks are then established for unified and systematic design, estimation, and uncertainty analysis in larger scale applications. The proposed algorithms' efficiency has been validated through comprehensive experiments. Specifically, the speed, accuracy and robustness of the fast plane extraction and the marker registration have been demonstrated to be superior to existing state-of-the-art algorithms. These algorithms have also been implemented in two groups of ARC applications to demonstrate the proposed framework's effectiveness, wherein the applications themselves have significant social and economic value. The first group is related to in-situ robotic machinery, including an autonomous manipulator for assembling digital architecture designs on construction sites to help improve productivity and quality; and an intelligent guidance and monitoring system for articulated machinery such as excavators to help improve safety. The second group emphasizes human-machine interaction to make ARC more effective, including a mobile Building Information Modeling and way-finding platform with discrete location recognition to increase indoor facility management efficiency; and a 3D scanning and modeling solution for rapid and cost-efficient dimension checking and concise as-built modeling.PHDCivil EngineeringUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/113481/1/cforrest_1.pd

    How significant are differences obtained by neglecting correlations when testing for deformation: A real case study using bootstrapping with terrestrial laser scanner observations approximated by b-spline surfaces

    Get PDF
    B-spline surfaces possess attractive properties such as a high degree of continuity or the local support of their basis functions. One of the major applications of B-spline surfaces in engineering geodesy is the least-square (LS) fitting of surfaces from, e.g., 3D point clouds obtained from terrestrial laser scanners (TLS). Such mathematical approximations allow one to test rigorously with a given significance level the deformation magnitude between point clouds taken at different epochs. Indeed, statistical tests cannot be applied when point clouds are processed in commonly used software such as CloudCompare, which restrict the analysis of deformation to simple deformation maps based on distance computation. For a trustworthy test decision and a resulting risk management, the stochastic model of the underlying observations needs, however, to be optimally specified. Since B-spline surface approximations necessitate Cartesian coordinates of the TLS observations, the diagonal variance covariance matrix (VCM) of the raw TLS measurements has to be transformed by means of the error propagation law. Unfortunately, this procedure induces mathematical correlations, which can strongly affect the chosen test statistics to analyse deformation, if neglected. This may lead potentially to rejecting wrongly the null hypothesis of no-deformation, with risky and expensive consequences. In this contribution, we propose to investigate the impact of mathematical correlations on test statistics, using real TLS observations from a bridge under load. As besides TLS, a highly precise laser tracker (LT) was used, the significance of the difference of the test statistics when the stochastic model is misspecified can be assessed. However, the underlying test distribution is hardly tractable so that only an adapted bootstrapping allows the computation of trustworthy p-values. Consecutively, the extent to which heteroscedasticity and mathematical correlations can be neglected or simplified without impacting the test decision is shown in a rigorous way, paving the way for a simplification based on the intensity model

    Fifth Annual Flight Mechanics/Estimation Theory Symposium

    Get PDF
    Various aspects of astrodynamics are considered including orbit calculations and trajectory determination. Other topics dealing with remote sensing systems, satellite navigation, and attitude control are included

    On the impact of correlations on the congruence test: a bootstrap approach: Case study: B-spline surface fitting from TLS observations

    Get PDF
    The detection of deformation is one of the major tasks in surveying engineering. It is meaningful only if the statistical significance of the distortions is correctly investigated, which often underlies a parametric modelization of the object under consideration. So-called regression B-spline approximation can be performed for point clouds of terrestrial laser scanners, allowing the setting of a specific congruence test based on the B-spline surfaces. Such tests are known to be strongly influenced by the underlying stochastic model chosen for the observation errors. The latter has to be correctly specified, which includes accounting for heteroscedasticity and correlations. In this contribution, we justify and make use of a parametric correlation model called the Matérn model to approximate the variance covariance matrix (VCM) of the residuals by performing their empirical mode decomposition. The VCM obtained is integrated into the computation of the congruence test statistics for a more trustworthy test decision. Using a real case study, we estimate the distribution of the test statistics with a bootstrap approach, where no parametric assumptions are made about the underlying population that generated the random sample. This procedure allows us to assess the impact of neglecting correlations on the critical value of the congruence test, highlighting their importance

    Aircraft state estimation using cameras and passive radar

    Get PDF
    Multiple target tracking (MTT) is a fundamental task in many application domains. It is a difficult problem to solve in general, so applications make use of domain specific and problem-specific knowledge to approach the problem by solving subtasks separately. This work puts forward a MTT framework (MTTF) which is based on the Bayesian recursive estimator (BRE). The MTTF extends a particle filter (PF) to handle the multiple targets and adds a probabilistic graphical model (PGM) data association stage to compute the mapping from detections to trackers. The MTTF was applied to the problem of passively monitoring airspace. Two applications were built: a passive radar MTT module and a comprehensive visual object tracking (VOT) system. Both applications require a solution to the MTT problem, for which the MTTF was utilized. The VOT system performed well on real data recorded at the University of Cape Town (UCT) as part of this investigation. The system was able to detect and track aircraft flying within the region of interest (ROI). The VOT system consisted of a single camera, an image processing module, the MTTF module and an evaluation module. The world coordinate frame target localization was within ±3.2 km and these results are presented on Google Earth. The image plane target localization has an average reprojection error of ±17.3 pixels. The VOT system achieved an average area under the curve value of 0.77 for all receiver operating characteristic curves. These performance figures are typical over the ±1 hr of video recordings taken from the UCT site. The passive radar application was tested on simulated data. The MTTF module was designed to connect to an existing passive radar system developed by Peralex Electronics Pty Ltd. The MTTF module estimated the number of targets in the scene and localized them within a 2D local world Cartesian coordinate system. The investigations encompass numerous areas of research as well as practical aspects of software engineering and systems design

    An evaluation of selected estimation methods for the processing of differential absorption lidar data

    Get PDF
    This work examines the application of selected estimation methods to path integrated direct detection CO₂ lidar data, with the objective of improving the precision in the estimates of the log power, and log power ratios. Particular emphasis is given to the optimal estimation techniques of Kalman filtering theory, and to the consequent requirements for system and measurement model identification. A dual wavelength system was designed and constructed, employing two hybridised TEA lasers, a co-axial transceiver, and direct detection.Over a period of several months, a database of differential absorption measurements was accumulated, each consisting of 10,000 dual wavelength lidar returns. Various wavelength pairs were used, including those recommended for the monitoring of H₂O, CO₂, NH₃ and C₂H₄. A subset of this database is used to evaluate the above mentioned estimation methods. The results are compared with simulated data files in which it was possible to control precisely process models which are believed to form an approximation to the real processes latent in the actual lidar data

    Risk Assessment for Alzheimer Patients, using GPS and Accelerometers with a Machine Learning Approach

    Get PDF
    Alzheimer is a pathology with an increasing incidence as people age. The epidemiology of the Alzheimer’s disease crosses every country at later stages in life and, for that motive, has become a major concern as expectancy of life is raising in developed world. The problem, as the disease progresses, becomes, the continuity of the person’s life as normal as possible and the insurance of safety and security for that person. It is known that Alzheimer patients tend to forget about important things such as their identity and location and for that it is important to provide that they become aware of such relevant information for them and for those who could provide them support. It is also known that people with Alzheimer tend to wander and, when that happens, they can get lost and become exposed to danger. The aim of this work is to assess if the person is getting away from usual paths and to monitor if the person falls which becomes riskier while wandering out of usual paths. The usage of GPS makes it possible to keep track of routes and, with the detection of possible deviations, it becomes possible to act accordingly, either issuing warnings for that person and later to carers and family. On the other hand, using machine learning to evaluate usual movements, it is possible to determine if a person falls or endures an excessively quiet position. With those strategies working together it is aimed to ensure safety of a person and request for assistance when high risk is assessed by the technological setup. To address these cases, the proposded setup will be based on a smartphone, together with a smartwatch, both carried by that person, as such devices already provide some needed sensors and GPS receiver while providing processing capabilities
    • …
    corecore