1,263 research outputs found
Expecting the Unexpected : Measuring Uncertainties in Mobile Robot Path Planning in Dynamic Envionments
Unexpected obstacles pose significant challenges to mobile robot navigation. In this paper we investigate how, based on the assumption that unexpected obstacles really follow patterns that can be exploited, a mobile robot can learn the locations within an environment that are likely to contain obstacles, and so plan optimal paths by avoiding these locations in subsequent navigation tasks. We propose the DUNC (Dynamically Updating Navigational Confidence) method to do this. We evaluate the performance of the DUNC method by comparing it with existing methods in a large number of randomly generated simulated test environments. our evaluations show that, by learning the likely locations of unexpected obstacles, the DUNC method can plan more efficient paths than existing approaches to this problem
A Machine Learning Trainable Model to Assess the Accuracy of Probabilistic Record Linkage
Record linkage (RL) is the process of identifying and linking data that relates to the same physical entity across multiple heterogeneous data sources. Deterministic linkage methods rely on the presence of common uniquely identifying attributes across all sources while probabilistic approaches use non-unique attributes and calculates similarity indexes for pair wise comparisons. A key component of record linkage is accuracy assessment â the process of manually verifying and validating matched pairs to further refine linkage parameters and increase its overall effectiveness. This process however is time-consuming and impractical when applied to large administrative data sources where millions of records must be linked. Additionally, it is potentially biased as the gold standard used is often the reviewerâs intuition. In this paper, we present an approach for assessing and refining the accuracy of probabilistic linkage based on different supervised machine learning methods (decision trees, naĂŻve Bayes, logistic regression, random forest, linear support vector machines and gradient boosted trees). We used data sets extracted from huge Brazilian socioeconomic and public health care data sources. These models were evaluated using receiver operating characteristic plots, sensitivity, specificity and positive predictive values collected from a 10-fold cross-validation method. Results show that logistic regression outperforms other classifiers and enables the creation of a generalized, very accurate model to validate linkage results
Within-Neighborhood Patterns and Sources of Particle Pollution: Mobile Monitoring and Geographic Information System Analysis in Four Communities in Accra, Ghana
BACKGROUND: Sources of air pollution in developing country cities include transportation and industrial pollution, biomass and coal fuel use, and resuspended dust from unpaved roads.
OBJECTIVES: Our goal was to understand within-neighborhood spatial variability of particulate matter (PM) in communities of varying socioeconomic status (SES) in Accra, Ghana, and to quantify the effects of nearby sources on local PM concentration.
METHODS: We conducted 1 week of morning and afternoon mobile and stationary air pollution measurements in four study neighborhoods. PM with aerodynamic diameters
RESULTS: In our measurement campaign, the geometric means of PM2.5 and PM10 along the mobile monitoring path were 21 and 49 microg/m3, respectively, in the neighborhood with highest SES and 39 and 96 microg/m3, respectively, in the neighborhood with lowest SES and highest population density. PM2.5 and PM10 were as high as 200 and 400 microg/m3, respectively, in some segments of the path. After adjusting for other factors, the factors that had the largest effects on local PM pollution were nearby wood and charcoal stoves, congested and heavy traffic, loose dirt road surface, and trash burning.
CONCLUSIONS: Biomass fuels, transportation, and unpaved roads may be important determinants of local PM variation in Accra neighborhoods. If confirmed by additional or supporting data, the results demonstrate the need for effective and equitable interventions and policies that reduce the impacts of traffic and biomass pollution
Performance of random forest when SNPs are in linkage disequilibrium
<p>Abstract</p> <p>Background</p> <p>Single nucleotide polymorphisms (SNPs) may be correlated due to linkage disequilibrium (LD). Association studies look for both direct and indirect associations with disease loci. In a Random Forest (RF) analysis, correlation between a true risk SNP and SNPs in LD may lead to diminished variable importance for the true risk SNP. One approach to address this problem is to select SNPs in linkage equilibrium (LE) for analysis. Here, we explore alternative methods for dealing with SNPs in LD: change the tree-building algorithm by building each tree in an RF only with SNPs in LE, modify the importance measure (IM), and use haplotypes instead of SNPs to build a RF.</p> <p>Results</p> <p>We evaluated the performance of our alternative methods by simulation of a spectrum of complex genetics models. When a haplotype rather than an individual SNP is the risk factor, we find that the original Random Forest method performed on SNPs provides good performance. When individual, genotyped SNPs are the risk factors, we find that the stronger the genetic effect, the stronger the effect LD has on the performance of the original RF. A revised importance measure used with the original RF is relatively robust to LD among SNPs; this revised importance measure used with the revised RF is sometimes inflated. Overall, we find that the revised importance measure used with the original RF is the best choice when the genetic model and the number of SNPs in LD with risk SNPs are unknown. For the haplotype-based method, under a multiplicative heterogeneity model, we observed a decrease in the performance of RF with increasing LD among the SNPs in the haplotype.</p> <p>Conclusion</p> <p>Our results suggest that by strategically revising the Random Forest method tree-building or importance measure calculation, power can increase when LD exists between SNPs. We conclude that the revised Random Forest method performed on SNPs offers an advantage of not requiring genotype phase, making it a viable tool for use in the context of thousands of SNPs, such as candidate gene studies and follow-up of top candidates from genome wide association studies.</p
Recommended from our members
A Search for MeV to TeV Neutrinos from Fast Radio Bursts with IceCube
We present two searches for IceCube neutrino events coincident with 28 fast radio bursts (FRBs) and 1 repeating FRB. The first improves on a previous IceCube analysis - searching for spatial and temporal correlation of events with FRBs at energies greater than roughly 50 GeV - by increasing the effective area by an order of magnitude. The second is a search for temporal correlation of MeV neutrino events with FRBs. No significant correlation is found in either search; therefore, we set upper limits on the time-integrated neutrino flux emitted by FRBs for a range of emission timescales less than one day. These are the first limits on FRB neutrino emission at the MeV scale, and the limits set at higher energies are an order-of-magnitude improvement over those set by any neutrino telescope
Recommended from our members
Efficient propagation of systematic uncertainties from calibration to analysis with the SnowStorm method in IceCube
Efficient treatment of systematic uncertainties that depend on a large number of nuisance parameters is a persistent difficulty in particle physics and astrophysics experiments. Where low-level effects are not amenable to simple parameterization or re-weighting, analyses often rely on discrete simulation sets to quantify the effects of nuisance parameters on key analysis observables. Such methods may become computationally untenable for analyses requiring high statistics Monte Carlo with a large number of nuisance degrees of freedom, especially in cases where these degrees of freedom parameterize the shape of a continuous distribution. In this paper we present a method for treating systematic uncertainties in a computationally efficient and comprehensive manner using a single simulation set with multiple and continuously varied nuisance parameters. This method is demonstrated for the case of the depth-dependent effective dust distribution within the IceCube Neutrino Telescope
Recommended from our members
Combined sensitivity to the neutrino mass ordering with JUNO, the IceCube Upgrade, and PINGU
The ordering of the neutrino mass eigenstates is one of the fundamental open questions in neutrino physics. While current-generation neutrino oscillation experiments are able to produce moderate indications on this ordering, upcoming experiments of the next generation aim to provide conclusive evidence. In this paper we study the combined performance of the two future multi-purpose neutrino oscillation experiments JUNO and the IceCube Upgrade, which employ two very distinct and complementary routes toward the neutrino mass ordering. The approach pursued by the 20 kt medium-baseline reactor neutrino experiment JUNO consists of a careful investigation of the energy spectrum of oscillated Îœe produced by ten nuclear reactor cores. The IceCube Upgrade, on the other hand, which consists of seven additional densely instrumented strings deployed in the center of IceCube DeepCore, will observe large numbers of atmospheric neutrinos that have undergone oscillations affected by Earth matter. In a joint fit with both approaches, tension occurs between their preferred mass-squared differences Îm312=m32-m12 within the wrong mass ordering. In the case of JUNO and the IceCube Upgrade, this allows to exclude the wrong ordering at >5Ï on a timescale of 3-7 years - even under circumstances that are unfavorable to the experiments' individual sensitivities. For PINGU, a 26-string detector array designed as a potential low-energy extension to IceCube, the inverted ordering could be excluded within 1.5 years (3 years for the normal ordering) in a joint analysis
Recommended from our members
Search for sources of astrophysical neutrinos using seven years of icecube cascade events
Low-background searches for astrophysical neutrino sources anywhere in the sky can be performed using cascade events induced by neutrinos of all flavors interacting in IceCube with energies as low as âŒ1 TeV. Previously we showed that, even with just two years of data, the resulting sensitivity to sources in the southern sky is competitive with IceCube and ANTARES analyses using muon tracks induced by charge current muon neutrino interactions - especially if the neutrino emission follows a soft energy spectrum or originates from an extended angular region. Here, we extend that work by adding five more years of data, significantly improving the cascade angular resolution, and including tests for point-like or diffuse Galactic emission to which this data set is particularly well suited. For many of the signal candidates considered, this analysis is the most sensitive of any experiment to date. No significant clustering was observed, and thus many of the resulting constraints are the most stringent to date. In this paper we will describe the improvements introduced in this analysis and discuss our results in the context of other recent work in neutrino astronomy
- âŠ