813 research outputs found
Improving the Creation of Hot Spot Policing Patrol Routes: Comparing Cognitive Heuristic Performance to an Automated Spatial Computation Approach
Hot spot policing involves the deployment of police patrols to places where high levels of crime have previously concentrated. The creation of patrol routes in these hot spots is mainly a manual process that involves using the results from an analysis of spatial patterns of crime to identify the areas and draw the routes that police officers are required to patrol. In this article we introduce a computational approach for automating the creation of hot spot policing patrol routes. The computational techniques we introduce created patrol routes that covered areas of higher levels of crime than an equivalent manual approach for creating hot spot policing patrol routes, and were more efficient in how they covered crime hot spots. Although the evidence on hot spot policing interventions shows they are effective in decreasing crime, the findings from the current research suggest that the impact of these interventions can potentially be greater when using the computational approaches that we introduce for creating hot spot policing patrol routes
Determinants of the income velocity of money in Portugal: 1891â1998
This paper performs a long-run time series analysis of the behaviour of the income velocity of money in Portugal between 1891 and 1998 by assessing the importance of both macroeconomic and institutional factors and looking for particularities in the Portuguese case. We estimate two cointegration vectors for the income velocity of money, macroeconomic variables and institutional variables. It is apparent that one of these vectors reflects the relationship between income velocity and macroeconomic variables, while the other reflects the relationship between income velocity and institutional variables. Moreover, a regression analysis reveals that the usual U-shaped pattern is displayed with a relatively late inflection point located around 1970, which is consistent with the Spanish case. It is further noted that this is a feature of countries with a late economic and institutional development process.info:eu-repo/semantics/publishedVersio
Evaluation of qPCR-Based Assays for Leprosy Diagnosis Directly in Clinical Specimens
The increased reliability and efficiency of the quantitative polymerase chain reaction (qPCR) makes it a promising tool for performing large-scale screening for infectious disease among high-risk individuals. To date, no study has evaluated the specificity and sensitivity of different qPCR assays for leprosy diagnosis using a range of clinical samples that could bias molecular results such as difficult-to-diagnose cases. In this study, qPCR assays amplifying different M. leprae gene targets, sodA, 16S rRNA, RLEP and Ag 85B were compared for leprosy differential diagnosis. qPCR assays were performed on frozen skin biopsy samples from a total of 62 patients: 21 untreated multibacillary (MB), 26 untreated paucibacillary (PB) leprosy patients, as well as 10 patients suffering from other dermatological diseases and 5 healthy donors. To develop standardized protocols and to overcome the bias resulted from using chromosome count cutoffs arbitrarily defined for different assays, decision tree classifiers were used to estimate optimum cutoffs and to evaluate the assays. As a result, we found a decreasing sensitivity for Ag 85B (66.1%), 16S rRNA (62.9%), and sodA (59.7%) optimized assay classifiers, but with similar maximum specificity for leprosy diagnosis. Conversely, the RLEP assay showed to be the most sensitive (87.1%). Moreover, RLEP assay was positive for 3 samples of patients originally not diagnosed as having leprosy, but these patients developed leprosy 5â10 years after the collection of the biopsy. In addition, 4 other samples of patients clinically classified as non-leprosy presented detectable chromosome counts in their samples by the RLEP assay suggesting that those patients either had leprosy that was misdiagnosed or a subclinical state of leprosy. Overall, these results are encouraging and suggest that RLEP assay could be useful as a sensitive diagnostic test to detect M. leprae infection before major clinical manifestations
Brane-World Gravity
The observable universe could be a 1+3-surface (the "brane") embedded in a
1+3+\textit{d}-dimensional spacetime (the "bulk"), with Standard Model
particles and fields trapped on the brane while gravity is free to access the
bulk. At least one of the \textit{d} extra spatial dimensions could be very
large relative to the Planck scale, which lowers the fundamental gravity scale,
possibly even down to the electroweak ( TeV) level. This revolutionary
picture arises in the framework of recent developments in M theory. The
1+10-dimensional M theory encompasses the known 1+9-dimensional superstring
theories, and is widely considered to be a promising potential route to quantum
gravity. At low energies, gravity is localized at the brane and general
relativity is recovered, but at high energies gravity "leaks" into the bulk,
behaving in a truly higher-dimensional way. This introduces significant changes
to gravitational dynamics and perturbations, with interesting and potentially
testable implications for high-energy astrophysics, black holes, and cosmology.
Brane-world models offer a phenomenological way to test some of the novel
predictions and corrections to general relativity that are implied by M theory.
This review analyzes the geometry, dynamics and perturbations of simple
brane-world models for cosmology and astrophysics, mainly focusing on warped
5-dimensional brane-worlds based on the Randall--Sundrum models. We also cover
the simplest brane-world models in which 4-dimensional gravity on the brane is
modified at \emph{low} energies -- the 5-dimensional Dvali--Gabadadze--Porrati
models. Then we discuss co-dimension two branes in 6-dimensional models.Comment: A major update of Living Reviews in Relativity 7:7 (2004)
"Brane-World Gravity", 119 pages, 28 figures, the update contains new
material on RS perturbations, including full numerical solutions of
gravitational waves and scalar perturbations, on DGP models, and also on 6D
models. A published version in Living Reviews in Relativit
Utilisation of an operative difficulty grading scale for laparoscopic cholecystectomy
Background
A reliable system for grading operative difficulty of laparoscopic cholecystectomy would standardise description of findings and reporting of outcomes. The aim of this study was to validate a difficulty grading system (Nassar scale), testing its applicability and consistency in two large prospective datasets.
Methods
Patient and disease-related variables and 30-day outcomes were identified in two prospective cholecystectomy databases: the multi-centre prospective cohort of 8820 patients from the recent CholeS Study and the single-surgeon series containing 4089 patients. Operative data and patient outcomes were correlated with Nassar operative difficultly scale, using Kendallâs tau for dichotomous variables, or JonckheereâTerpstra tests for continuous variables. A ROC curve analysis was performed, to quantify the predictive accuracy of the scale for each outcome, with continuous outcomes dichotomised, prior to analysis.
Results
A higher operative difficulty grade was consistently associated with worse outcomes for the patients in both the reference and CholeS cohorts. The median length of stay increased from 0 to 4 days, and the 30-day complication rate from 7.6 to 24.4% as the difficulty grade increased from 1 to 4/5 (both pâ<â0.001). In the CholeS cohort, a higher difficulty grade was found to be most strongly associated with conversion to open and 30-day mortality (AUROCâ=â0.903, 0.822, respectively). On multivariable analysis, the Nassar operative difficultly scale was found to be a significant independent predictor of operative duration, conversion to open surgery, 30-day complications and 30-day reintervention (all pâ<â0.001).
Conclusion
We have shown that an operative difficulty scale can standardise the description of operative findings by multiple grades of surgeons to facilitate audit, training assessment and research. It provides a tool for reporting operative findings, disease severity and technical difficulty and can be utilised in future research to reliably compare outcomes according to case mix and intra-operative difficulty
Observation of associated near-side and away-side long-range correlations in âsNN=5.02ââTeV proton-lead collisions with the ATLAS detector
Two-particle correlations in relative azimuthal angle (ÎÏ) and pseudorapidity (Îη) are measured in âsNN=5.02ââTeV p+Pb collisions using the ATLAS detector at the LHC. The measurements are performed using approximately 1ââÎŒb-1 of data as a function of transverse momentum (pT) and the transverse energy (ÎŁETPb) summed over 3.1<η<4.9 in the direction of the Pb beam. The correlation function, constructed from charged particles, exhibits a long-range (2<|Îη|<5) ânear-sideâ (ÎÏâŒ0) correlation that grows rapidly with increasing ÎŁETPb. A long-range âaway-sideâ (ÎÏâŒÏ) correlation, obtained by subtracting the expected contributions from recoiling dijets and other sources estimated using events with small ÎŁETPb, is found to match the near-side correlation in magnitude, shape (in Îη and ÎÏ) and ÎŁETPb dependence. The resultant ÎÏ correlation is approximately symmetric about Ï/2, and is consistent with a dominant cosâĄ2ÎÏ modulation for all ÎŁETPb ranges and particle pT
- âŠ