408 research outputs found
Network analysis of a corpus of undeciphered Indus civilization inscriptions indicates syntactic organization
Archaeological excavations in the sites of the Indus Valley civilization
(2500-1900 BCE) in Pakistan and northwestern India have unearthed a large
number of artifacts with inscriptions made up of hundreds of distinct signs. To
date there is no generally accepted decipherment of these sign sequences and
there have been suggestions that the signs could be non-linguistic. Here we
apply complex network analysis techniques to a database of available Indus
inscriptions, with the aim of detecting patterns indicative of syntactic
organization. Our results show the presence of patterns, e.g., recursive
structures in the segmentation trees of the sequences, that suggest the
existence of a grammar underlying these inscriptions.Comment: 17 pages (includes 4 page appendix containing Indus sign list), 14
figure
The modern pollen-vegetation relationship of a tropical forest-savannah mosaic landscape, Ghana, West Africa
Transitions between forest and savannah vegetation types in fossil pollen records are often poorly understood due to over-production by taxa such as Poaceae and a lack of modern pollen-vegetation studies. Here, modern pollen assemblages from within a forest-savannah transition in West Africa are presented and compared, their characteristic taxa discussed, and implications for the fossil record considered. Fifteen artificial pollen traps were deployed for 1 year, to collect pollen rain from three vegetation plots within the forest-savannah transition in Ghana. High percentages of Poaceae and Melastomataceae/Combretaceae were recorded in all three plots. Erythrophleum suaveolens characterised the forest plot, Manilkara obovata the transition plot and Terminalia the savannah plot. The results indicate that Poaceae pollen influx rates provide the best representation of the forest-savannah gradient, and that a Poaceae abundance of >40% should be considered as indicative of savannah-type vegetation in the fossil record
How khipus indicated labour contributions in an Andean village: an explanation of colour banding, seriation and ethnocategories
This research was supported by a Global Exploration Grant from the National Geographic Society (GEFNE120-14).New archival and ethnographic evidence reveals that Inka style khipus were used in the Andean community of Santiago de Anchucaya to record contributions to communal labour obligations until the 1940s. Archival testimony from the last khipu specialist in Anchucaya, supplemented by interviews with his grandson, provides the first known expert explanation for how goods, labour obligations, and social groups were indicated on Inka style Andean khipus. This evidence, combined with the analysis of Anchucaya khipus in the Museo Nacional de ArqueologĂa, AntropologĂa y Historia Peruana, furnishes a local model for the relationship between the two most frequent colour patterns (colour banding and seriation) that occur in khipus. In this model, colour banding is associated with individual data whilst seriation is associated with aggregated data. The archival and ethnographic evidence also explains how labour and goods were categorized in uniquely Andean ways as they were represented on khipus.PostprintPeer reviewe
DFT-inspired methods for quantum thermodynamics
In the framework of quantum thermodynamics, we propose a method to
quantitatively describe thermodynamic quantities for out-of-equilibrium
interacting many-body systems. The method is articulated in various
approximation protocols which allow to achieve increasing levels of accuracy,
it is relatively simple to implement even for medium and large number of
interactive particles, and uses tools and concepts from density functional
theory. We test the method on the driven Hubbard dimer at half filling, and
compare exact and approximate results. We show that the proposed method
reproduces the average quantum work to high accuracy: for a very large region
of parameter space (which cuts across all dynamical regimes) estimates are
within 10% of the exact results
Cluster Lenses
Clusters of galaxies are the most recently assembled, massive, bound
structures in the Universe. As predicted by General Relativity, given their
masses, clusters strongly deform space-time in their vicinity. Clusters act as
some of the most powerful gravitational lenses in the Universe. Light rays
traversing through clusters from distant sources are hence deflected, and the
resulting images of these distant objects therefore appear distorted and
magnified. Lensing by clusters occurs in two regimes, each with unique
observational signatures. The strong lensing regime is characterized by effects
readily seen by eye, namely, the production of giant arcs, multiple-images, and
arclets. The weak lensing regime is characterized by small deformations in the
shapes of background galaxies only detectable statistically. Cluster lenses
have been exploited successfully to address several important current questions
in cosmology: (i) the study of the lens(es) - understanding cluster mass
distributions and issues pertaining to cluster formation and evolution, as well
as constraining the nature of dark matter; (ii) the study of the lensed objects
- probing the properties of the background lensed galaxy population - which is
statistically at higher redshifts and of lower intrinsic luminosity thus
enabling the probing of galaxy formation at the earliest times right up to the
Dark Ages; and (iii) the study of the geometry of the Universe - as the
strength of lensing depends on the ratios of angular diameter distances between
the lens, source and observer, lens deflections are sensitive to the value of
cosmological parameters and offer a powerful geometric tool to probe Dark
Energy. In this review, we present the basics of cluster lensing and provide a
current status report of the field.Comment: About 120 pages - Published in Open Access at:
http://www.springerlink.com/content/j183018170485723/ . arXiv admin note:
text overlap with arXiv:astro-ph/0504478 and arXiv:1003.3674 by other author
Simple large wood structures promote hydromorphological heterogeneity and benthic macroinvertebrate diversity in low-gradient rivers
This work has been carried out within the SMART Joint Doctorate Programme âScience for the MAnagement of Rivers and their Tidal systemsâ funded by the Erasmus Mundus programme of the European Union
Utilisation of an operative difficulty grading scale for laparoscopic cholecystectomy
Background
A reliable system for grading operative difficulty of laparoscopic cholecystectomy would standardise description of findings and reporting of outcomes. The aim of this study was to validate a difficulty grading system (Nassar scale), testing its applicability and consistency in two large prospective datasets.
Methods
Patient and disease-related variables and 30-day outcomes were identified in two prospective cholecystectomy databases: the multi-centre prospective cohort of 8820 patients from the recent CholeS Study and the single-surgeon series containing 4089 patients. Operative data and patient outcomes were correlated with Nassar operative difficultly scale, using Kendallâs tau for dichotomous variables, or JonckheereâTerpstra tests for continuous variables. A ROC curve analysis was performed, to quantify the predictive accuracy of the scale for each outcome, with continuous outcomes dichotomised, prior to analysis.
Results
A higher operative difficulty grade was consistently associated with worse outcomes for the patients in both the reference and CholeS cohorts. The median length of stay increased from 0 to 4 days, and the 30-day complication rate from 7.6 to 24.4% as the difficulty grade increased from 1 to 4/5 (both pâ<â0.001). In the CholeS cohort, a higher difficulty grade was found to be most strongly associated with conversion to open and 30-day mortality (AUROCâ=â0.903, 0.822, respectively). On multivariable analysis, the Nassar operative difficultly scale was found to be a significant independent predictor of operative duration, conversion to open surgery, 30-day complications and 30-day reintervention (all pâ<â0.001).
Conclusion
We have shown that an operative difficulty scale can standardise the description of operative findings by multiple grades of surgeons to facilitate audit, training assessment and research. It provides a tool for reporting operative findings, disease severity and technical difficulty and can be utilised in future research to reliably compare outcomes according to case mix and intra-operative difficulty
Jet energy measurement with the ATLAS detector in proton-proton collisions at root s=7 TeV
The jet energy scale and its systematic uncertainty are determined for jets measured with the ATLAS detector at the LHC in proton-proton collision data at a centre-of-mass energy of âs = 7TeV corresponding to an integrated luminosity of 38 pb-1. Jets are reconstructed with the anti-kt algorithm with distance parameters R=0. 4 or R=0. 6. Jet energy and angle corrections are determined from Monte Carlo simulations to calibrate jets with transverse momenta pTâ„20 GeV and pseudorapidities {pipe}η{pipe}<4. 5. The jet energy systematic uncertainty is estimated using the single isolated hadron response measured in situ and in test-beams, exploiting the transverse momentum balance between central and forward jets in events with dijet topologies and studying systematic variations in Monte Carlo simulations. The jet energy uncertainty is less than 2. 5 % in the central calorimeter region ({pipe}η{pipe}<0. 8) for jets with 60â€pT<800 GeV, and is maximally 14 % for pT<30 GeV in the most forward region 3. 2â€{pipe}η{pipe}<4. 5. The jet energy is validated for jet transverse momenta up to 1 TeV to the level of a few percent using several in situ techniques by comparing a well-known reference such as the recoiling photon pT, the sum of the transverse momenta of tracks associated to the jet, or a system of low-pT jets recoiling against a high-pT jet. More sophisticated jet calibration schemes are presented based on calorimeter cell energy density weighting or hadronic properties of jets, aiming for an improved jet energy resolution and a reduced flavour dependence of the jet response. The systematic uncertainty of the jet energy determined from a combination of in situ techniques is consistent with the one derived from single hadron response measurements over a wide kinematic range. The nominal corrections and uncertainties are derived for isolated jets in an inclusive sample of high-pT jets. Special cases such as event topologies with close-by jets, or selections of samples with an enhanced content of jets originating from light quarks, heavy quarks or gluons are also discussed and the corresponding uncertainties are determined. © 2013 CERN for the benefit of the ATLAS collaboration
Measurement of the inclusive and dijet cross-sections of b-jets in pp collisions at sqrt(s) = 7 TeV with the ATLAS detector
The inclusive and dijet production cross-sections have been measured for jets
containing b-hadrons (b-jets) in proton-proton collisions at a centre-of-mass
energy of sqrt(s) = 7 TeV, using the ATLAS detector at the LHC. The
measurements use data corresponding to an integrated luminosity of 34 pb^-1.
The b-jets are identified using either a lifetime-based method, where secondary
decay vertices of b-hadrons in jets are reconstructed using information from
the tracking detectors, or a muon-based method where the presence of a muon is
used to identify semileptonic decays of b-hadrons inside jets. The inclusive
b-jet cross-section is measured as a function of transverse momentum in the
range 20 < pT < 400 GeV and rapidity in the range |y| < 2.1. The bbbar-dijet
cross-section is measured as a function of the dijet invariant mass in the
range 110 < m_jj < 760 GeV, the azimuthal angle difference between the two jets
and the angular variable chi in two dijet mass regions. The results are
compared with next-to-leading-order QCD predictions. Good agreement is observed
between the measured cross-sections and the predictions obtained using POWHEG +
Pythia. MC@NLO + Herwig shows good agreement with the measured bbbar-dijet
cross-section. However, it does not reproduce the measured inclusive
cross-section well, particularly for central b-jets with large transverse
momenta.Comment: 10 pages plus author list (21 pages total), 8 figures, 1 table, final
version published in European Physical Journal
A BAC pooling strategy combined with PCR-based screenings in a large, highly repetitive genome enables integration of the maize genetic and physical maps
BACKGROUND: Molecular markers serve three important functions in physical map assembly. First, they provide anchor points to genetic maps facilitating functional genomic studies. Second, they reduce the overlap required for BAC contig assembly from 80 to 50 percent. Finally, they validate assemblies based solely on BAC fingerprints. We employed a six-dimensional BAC pooling strategy in combination with a high-throughput PCR-based screening method to anchor the maize genetic and physical maps. RESULTS: A total of 110,592 maize BAC clones (~ 6x haploid genome equivalents) were pooled into six different matrices, each containing 48 pools of BAC DNA. The quality of the BAC DNA pools and their utility for identifying BACs containing target genomic sequences was tested using 254 PCR-based STS markers. Five types of PCR-based STS markers were screened to assess potential uses for the BAC pools. An average of 4.68 BAC clones were identified per marker analyzed. These results were integrated with BAC fingerprint data generated by the Arizona Genomics Institute (AGI) and the Arizona Genomics Computational Laboratory (AGCoL) to assemble the BAC contigs using the FingerPrinted Contigs (FPC) software and contribute to the construction and anchoring of the physical map. A total of 234 markers (92.5%) anchored BAC contigs to their genetic map positions. The results can be viewed on the integrated map of maize [1,2]. CONCLUSION: This BAC pooling strategy is a rapid, cost effective method for genome assembly and anchoring. The requirement for six replicate positive amplifications makes this a robust method for use in large genomes with high amounts of repetitive DNA such as maize. This strategy can be used to physically map duplicate loci, provide order information for loci in a small genetic interval or with no genetic recombination, and loci with conflicting hybridization-based information
- âŠ