4,049,959 research outputs found
Classification accuracy increase using multisensor data fusion
The practical use of very high resolution visible and near-infrared (VNIR) data is still growing (IKONOS, Quickbird, GeoEye-1, etc.)
but for classification purposes the number of bands is limited in comparison to full spectral imaging. These limitations may lead to the
confusion of materials such as different roofs, pavements, roads, etc. and therefore may provide wrong interpretation and use of classification
products. Employment of hyperspectral data is another solution, but their low spatial resolution (comparing to multispectral
data) restrict their usage for many applications. Another improvement can be achieved by fusion approaches of multisensory data since
this may increase the quality of scene classification. Integration of Synthetic Aperture Radar (SAR) and optical data is widely performed
for automatic classification, interpretation, and change detection. In this paper we present an approach for very high resolution
SAR and multispectral data fusion for automatic classification in urban areas. Single polarization TerraSAR-X (SpotLight mode) and
multispectral data are integrated using the INFOFUSE framework, consisting of feature extraction (information fission), unsupervised
clustering (data representation on a finite domain and dimensionality reduction), and data aggregation (Bayesian or neural network).
This framework allows a relevant way of multisource data combination following consensus theory. The classification is not influenced
by the limitations of dimensionality, and the calculation complexity primarily depends on the step of dimensionality reduction. Fusion
of single polarization TerraSAR-X, WorldView-2 (VNIR or full set), and Digital Surface Model (DSM) data allow for different types
of urban objects to be classified into predefined classes of interest with increased accuracy. The comparison to classification results
of WorldView-2 multispectral data (8 spectral bands) is provided and the numerical evaluation of the method in comparison to other
established methods illustrates the advantage in the classification accuracy for many classes such as buildings, low vegetation, sport
objects, forest, roads, rail roads, etc
Accuracy of areal interpolation methods for count data
The combination of several socio-economic data bases originating from
different administrative sources collected on several different partitions of a
geographic zone of interest into administrative units induces the so called
areal interpolation problem. This problem is that of allocating the data from a
set of source spatial units to a set of target spatial units. A particular case
of that problem is the re-allocation to a single target partition which is a
regular grid. At the European level for example, the EU directive 'INSPIRE', or
INfrastructure for SPatial InfoRmation, encourages the states to provide
socio-economic data on a common grid to facilitate economic studies across
states. In the literature, there are three main types of such techniques:
proportional weighting schemes, smoothing techniques and regression based
interpolation. We propose a stochastic model based on Poisson point patterns to
study the statistical accuracy of these techniques for regular grid targets in
the case of count data. The error depends on the nature of the target variable
and its correlation with the auxiliary variable. For simplicity, we restrict
attention to proportional weighting schemes and Poisson regression based
methods. Our conclusion is that there is no technique which always dominates
Model waveform accuracy standards for gravitational wave data analysis
Model waveforms are used in gravitational wave data analysis to detect and then to measure the properties of a source by matching the model waveforms to the signal from a detector. This paper derives accuracy standards for model waveforms which are sufficient to ensure that these data analysis applications are capable of extracting the full scientific content of the data, but without demanding excessive accuracy that would place undue burdens on the model waveform simulation community. These accuracy standards are intended primarily for broadband model waveforms produced by numerical simulations, but the standards are quite general and apply equally to such waveforms produced by analytical or hybrid analytical-numerical methods
Recovering from Biased Data: Can Fairness Constraints Improve Accuracy?
Multiple fairness constraints have been proposed in the literature, motivated by a range of concerns about how demographic groups might be treated unfairly by machine learning classifiers. In this work we consider a different motivation; learning from biased training data. We posit several ways in which training data may be biased, including having a more noisy or negatively biased labeling process on members of a disadvantaged group, or a decreased prevalence of positive or negative examples from the disadvantaged group, or both. Given such biased training data, Empirical Risk Minimization (ERM) may produce a classifier that not only is biased but also has suboptimal accuracy on the true data distribution. We examine the ability of fairness-constrained ERM to correct this problem. In particular, we find that the Equal Opportunity fairness constraint [Hardt et al., 2016] combined with ERM will provably recover the Bayes optimal classifier under a range of bias models. We also consider other recovery methods including re-weighting the training data, Equalized Odds, and Demographic Parity, and Calibration. These theoretical results provide additional motivation for considering fairness interventions even if an actor cares primarily about accuracy
UK open source crime data: accuracy and possibilities for research
In the United Kingdom, since 2011 data regarding individual police recorded crimes have been made openly available to the public via the police.uk website. To protect the location privacy of victims these data are obfuscated using geomasking techniques to reduce their spatial accuracy. This paper examines the spatial accuracy of the police.uk data to determine at what level(s) of spatial resolution – if any – it is suitable for analysis in the context of theory testing and falsification, evaluation research, or crime analysis. Police.uk data are compared to police recorded data for one large metropolitan Police Force and spatial accuracy is quantified for four different levels of geography across five crime types. Hypotheses regarding systematic errors are tested using appropriate statistical approaches, including methods of maximum likelihood. Finally, a “best-fit” statistical model is presented to explain the error as well as to develop a model that can correct it. The implications of the findings for researchers using the police.uk data for spatial analysis are discussed
- …
