4,049,959 research outputs found

    Classification accuracy increase using multisensor data fusion

    Get PDF
    The practical use of very high resolution visible and near-infrared (VNIR) data is still growing (IKONOS, Quickbird, GeoEye-1, etc.) but for classification purposes the number of bands is limited in comparison to full spectral imaging. These limitations may lead to the confusion of materials such as different roofs, pavements, roads, etc. and therefore may provide wrong interpretation and use of classification products. Employment of hyperspectral data is another solution, but their low spatial resolution (comparing to multispectral data) restrict their usage for many applications. Another improvement can be achieved by fusion approaches of multisensory data since this may increase the quality of scene classification. Integration of Synthetic Aperture Radar (SAR) and optical data is widely performed for automatic classification, interpretation, and change detection. In this paper we present an approach for very high resolution SAR and multispectral data fusion for automatic classification in urban areas. Single polarization TerraSAR-X (SpotLight mode) and multispectral data are integrated using the INFOFUSE framework, consisting of feature extraction (information fission), unsupervised clustering (data representation on a finite domain and dimensionality reduction), and data aggregation (Bayesian or neural network). This framework allows a relevant way of multisource data combination following consensus theory. The classification is not influenced by the limitations of dimensionality, and the calculation complexity primarily depends on the step of dimensionality reduction. Fusion of single polarization TerraSAR-X, WorldView-2 (VNIR or full set), and Digital Surface Model (DSM) data allow for different types of urban objects to be classified into predefined classes of interest with increased accuracy. The comparison to classification results of WorldView-2 multispectral data (8 spectral bands) is provided and the numerical evaluation of the method in comparison to other established methods illustrates the advantage in the classification accuracy for many classes such as buildings, low vegetation, sport objects, forest, roads, rail roads, etc

    Accuracy of areal interpolation methods for count data

    Full text link
    The combination of several socio-economic data bases originating from different administrative sources collected on several different partitions of a geographic zone of interest into administrative units induces the so called areal interpolation problem. This problem is that of allocating the data from a set of source spatial units to a set of target spatial units. A particular case of that problem is the re-allocation to a single target partition which is a regular grid. At the European level for example, the EU directive 'INSPIRE', or INfrastructure for SPatial InfoRmation, encourages the states to provide socio-economic data on a common grid to facilitate economic studies across states. In the literature, there are three main types of such techniques: proportional weighting schemes, smoothing techniques and regression based interpolation. We propose a stochastic model based on Poisson point patterns to study the statistical accuracy of these techniques for regular grid targets in the case of count data. The error depends on the nature of the target variable and its correlation with the auxiliary variable. For simplicity, we restrict attention to proportional weighting schemes and Poisson regression based methods. Our conclusion is that there is no technique which always dominates

    Model waveform accuracy standards for gravitational wave data analysis

    Get PDF
    Model waveforms are used in gravitational wave data analysis to detect and then to measure the properties of a source by matching the model waveforms to the signal from a detector. This paper derives accuracy standards for model waveforms which are sufficient to ensure that these data analysis applications are capable of extracting the full scientific content of the data, but without demanding excessive accuracy that would place undue burdens on the model waveform simulation community. These accuracy standards are intended primarily for broadband model waveforms produced by numerical simulations, but the standards are quite general and apply equally to such waveforms produced by analytical or hybrid analytical-numerical methods

    Recovering from Biased Data: Can Fairness Constraints Improve Accuracy?

    Get PDF
    Multiple fairness constraints have been proposed in the literature, motivated by a range of concerns about how demographic groups might be treated unfairly by machine learning classifiers. In this work we consider a different motivation; learning from biased training data. We posit several ways in which training data may be biased, including having a more noisy or negatively biased labeling process on members of a disadvantaged group, or a decreased prevalence of positive or negative examples from the disadvantaged group, or both. Given such biased training data, Empirical Risk Minimization (ERM) may produce a classifier that not only is biased but also has suboptimal accuracy on the true data distribution. We examine the ability of fairness-constrained ERM to correct this problem. In particular, we find that the Equal Opportunity fairness constraint [Hardt et al., 2016] combined with ERM will provably recover the Bayes optimal classifier under a range of bias models. We also consider other recovery methods including re-weighting the training data, Equalized Odds, and Demographic Parity, and Calibration. These theoretical results provide additional motivation for considering fairness interventions even if an actor cares primarily about accuracy

    UK open source crime data: accuracy and possibilities for research

    Get PDF
    In the United Kingdom, since 2011 data regarding individual police recorded crimes have been made openly available to the public via the police.uk website. To protect the location privacy of victims these data are obfuscated using geomasking techniques to reduce their spatial accuracy. This paper examines the spatial accuracy of the police.uk data to determine at what level(s) of spatial resolution – if any – it is suitable for analysis in the context of theory testing and falsification, evaluation research, or crime analysis. Police.uk data are compared to police recorded data for one large metropolitan Police Force and spatial accuracy is quantified for four different levels of geography across five crime types. Hypotheses regarding systematic errors are tested using appropriate statistical approaches, including methods of maximum likelihood. Finally, a “best-fit” statistical model is presented to explain the error as well as to develop a model that can correct it. The implications of the findings for researchers using the police.uk data for spatial analysis are discussed
    corecore