1,966 research outputs found
Cardiac re-synchronization therapy in a patient with isolated ventricular non-compaction: a case report.
Isolated ventricular non-compaction (IVNC) is a rare, congenital, unclassified cardiomyopathy characterized by prominent trabecular meshwork and deep recesses. Major clinical manifestations of IVNC are heart failure, atrial and ventricular arrhythmias, and thrombo-embolic events. We describe a case of a 69-year-old woman in whom the diagnosis of IVNC was discovered late, whereas former echocardiographic examinations were considered normal. She was known for systolic left ventricular dysfunction for 3 years and then became symptomatic (NYHA III). In the past, she suffered from multiple episodes of deep vein thrombosis and pulmonary embolism. Electrocardiogram revealed a wide QRS complex, and transthoracic echocardiography showed typical apical thickening of the left and right ventricular myocardial wall with two distinct layers. The ratio of non-compacted to compacted myocardium was >2:1. Cardiac MRI confirmed the echocardiographic images. Cerebral MRI revealed multiple ischaemic sequellae. In view of the persistent refractory, heart failure in medical treatment of patients with classical criteria for cardiac re-synchronization therapy, as well as the ventricular arrhythmias, a biventricular automatic intracardiac defibrillator (biventricular ICD) was implanted. The 2-year follow-up period was characterized by improvement of NYHA functional class from III to I and increasing in left ventricular function. We hereby present a case of IVNC with favourable outcome after biventricular ICD implantation. Cardiac re-synchronization therapy could be considered in the management of this pathology
Identification of Wax Esters in Latent Print Residues by Gas Chromatography-Mass Spectromertry and Their Potential Use as Aging Parameters
Recent studies show that the composition of fingerprint residue varies significantly from the same donor as well as between donors. This variability is a major drawback in latent print dating issues. This study aimed, therefore, at the definition of a parameter that is less variable from print to print, using a ratio of peak area of a target compound degrading over time divided by the summed area of peaks of more stable compounds also found in latent print residues.Gas chromatography-mass spectrometry (GC/MS) analysis of the initial lipid composition of latent prints identifies four main classes of compounds that can be used in the definition of an aging parameter: fatty acids, sterols, sterol precursors, and wax esters (WEs). Although the entities composing the first three groups are quite well known, those composing WEs are poorly reported. Therefore, the first step of the present work was to identify WE compounds present in latent print residues deposited by different donors. Of 29 WEs recorded in the chromatograms, seven were observed in the majority of samples.The identified WE compounds were subsequently used in the definition of ratios in combination with squalene and cholesterol to reduce the variability of the initial composition between latent print residues from different persons and more particularly from the same person. Finally, the influence of a latent print enhancement process on the initial composition was studied by analyzing traces after treatment with magnetic powder, 1,2-indanedione, and cyanoacrylate
The tension between fire risk and carbon storage: evaluating U.S. carbon and fire management strategies through ecosystem models
Fire risk and carbon storage are related environmental issues because fire reduction results in carbon storage through the buildup of woody vegetation, and stored carbon is a fuel for fires. The sustainability of the U.S. carbon sink and the extent of fire activity in the next 100 yr depend in part on the type and effectiveness of fire reduction employed. Previous studies have bracketed the range of dynamics from continued fire reduction to the complete failure of fire reduction activities. To improve these estimates, it is necessary to explicitly account for fire reduction in terrestrial models. A new fire reduction submodel that estimates the spatiotemporal pattern of reduction across the United States was developed using gridded data on biomass, climate, land-use, population, and economic factors. To the authors’ knowledge, it is the first large-scale, gridded fire model that explicitly accounts for fire reduction. The model was calibrated to 1° × 1° burned area statistics [Global Burnt Area 2000 Project (GBA-2000)] and compared favorably to three important diagnostics. The model was then implemented in a spatially explicit ecosystem model and used to analyze 1620 scenarios of future fire risk and fire reduction strategies. Under scenarios of climate change and urbanization, burned area and carbon emissions both increased in scenarios where fire reduction efforts were not adjusted to match new patterns of fire risk. Fuel reducing management strategies reduced burned area and fire risk, but also limited carbon storage. These results suggest that to promote carbon storage and minimize fire risk in the future, fire reduction efforts will need to be increased and spatially adjusted and will need to employ a mixture of fuel-reducing and non-fuel-reducing strategies
The Neutron Halo in Heavy Nuclei Calculated with the Gogny Force
The proton and neutron density distributions, one- and two-neutron separation
energies and radii of nuclei for which neutron halos are experimentally
observed, are calculated using the self-consistent Hartree-Fock-Bogoliubov
method with the effective interaction of Gogny. Halo factors are evaluated
assuming hydrogen-like antiproton wave functions. The factors agree well with
experimental data. They are close to those obtained with Skyrme forces and with
the relativistic mean field approach.Comment: 13 pages in Latex and 17 figures in ep
RLFC: Random Access Light Field Compression using Key Views and Bounded Integer Encoding
We present a new hierarchical compression scheme for encoding light field
images (LFI) that is suitable for interactive rendering. Our method (RLFC)
exploits redundancies in the light field images by constructing a tree
structure. The top level (root) of the tree captures the common high-level
details across the LFI, and other levels (children) of the tree capture
specific low-level details of the LFI. Our decompressing algorithm corresponds
to tree traversal operations and gathers the values stored at different levels
of the tree. Furthermore, we use bounded integer sequence encoding which
provides random access and fast hardware decoding for compressing the blocks of
children of the tree. We have evaluated our method for 4D two-plane
parameterized light fields. The compression rates vary from 0.08 - 2.5 bits per
pixel (bpp), resulting in compression ratios of around 200:1 to 20:1 for a PSNR
quality of 40 to 50 dB. The decompression times for decoding the blocks of LFI
are 1 - 3 microseconds per channel on an NVIDIA GTX-960 and we can render new
views with a resolution of 512X512 at 200 fps. Our overall scheme is simple to
implement and involves only bit manipulations and integer arithmetic
operations.Comment: Accepted for publication at Symposium on Interactive 3D Graphics and
Games (I3D '19
La datation des traces digitales (partie II): proposition d'une approche formelle
«Quel est l'âge de cette trace digitale?» Cette question est relativement souvent soulevée durant une investigation ou au tribunal, lorsque la personne suspectée admet avoir laissé ses traces sur une scène de crime mais prétend l'avoir fait à un autre moment que celui du crime et pour une raison innocente. La première partie de cet article mettait en évidence le manque de consensus actuel dans les réponses données à cette question par les experts du domaine, ainsi que le fait qu'aucune méthodologie n'est pour l'heure validée et acceptée par la communauté forensique. C'est pourquoi ce deuxième article propose une approche formelle et pragmatique afin d'aborder la question de la datation des traces digitales en se basant sur la recherche actuelle dans le domaine du vieillissement de composés lipidiques détectés dans les traces digitales. Cette approche permet ainsi d'identifier quel type d'information le scientifique serait capable d'apporter aux enquêteurs et/ou à la Cour lors de cas de datation des traces digitales à l'heure actuelle, dans quelles conditions, et quels sont les développements encore nécessaires
Aging of target lipid parameters in fingermark residue using GC/MS: effects of influence factors and perspectives for dating purposes
Despite the recurrence of fingermark dating issues and the research conducted on fingermark composition and aging, no dating methodology has yet been developed and validated. In order to further evaluate the possibility of developing dating methodologies based on the fingermark composition, this research proposed an in-depth study of the aging of target lipid parameters found in fingermark residue and exposed to different in fluence factors. The selected analytical technique was gas chromatography coupled with mass spectrometry (GC/MS). The effects of donor, substrate and enhancement techniques on the selected parameters were firstly evaluated. These factors were called known factors, as their value could be obtained in real caseworks. Using principal component analysis (PCA) and univariate exponential regression, this study highlighted the fact that the effects of these factors were larger than the aging effects, thus preventing the observation of relevant aging patterns. From a fingermark dating perspective, the specific value of these known factors should thus be included in aging models newly built for each case. Then, the effects of deposition moment, pressure, temperature and lighting were also evaluated. These factors were called unknown factors, as their specific value would never be precisely obtained in caseworks. Aging models should thus be particularly robust to their effects and for this reason, different chemometric tools were tested: PCA, univariate exponentialregression and partial least square regression(PLSR). While the first two models allowed observing interesting aging patterns regardless of the value of the applied influence factors, PLSR gave poorer results, as large deviations were obtained. Finally, in order to evaluate the potential of such modelling in realistic situations, blind analyses were carried out on eight test fingermarks. The age of five of them was correctly estimated using soft independent modelling of class analogy analysis (SIMCA) based on PCA classes, univariate exponential linear regression and PLSR. Furthermore, a probabilistic approach using the calculation of likelihood ratios (LR) through the construction of a Bayesian network was also tested. While the age of all test fingermarks were correctly evaluated when the storage conditions were known, the results were not significant when these conditions were unknown. Thus, this model clearly highlighted the impact of storage conditions on correct age evaluation. This research showed that reproducible aging modelling could be obtained based on fingermark residue exposed to influence factors, as well as promising age estimations. However, the proposed models are still not applicable in practice. Further studies should be conducted concerning the impact of influence factors (in particular, storage conditions) in order to precisely evaluate in which conditions significant evaluations could be obtained. Furthermore, these models should be properly validated before any application in real caseworks could be envisaged
High precision determination of the -evolution of the Bjorken Sum
We present a significantly improved determination of the Bjorken Sum for
0.64.8 GeV using precise new and
data taken with the CLAS detector at Jefferson Lab. A higher-twist analysis of
the -dependence of the Bjorken Sum yields the twist-4 coefficient
. This leads to the color
polarizabilities and
. The strong force coupling is determined to be
\alpha_{s}^{\overline{\mbox{ MS}}}(M_{Z}^{2})=0.1124\pm0.0061, which has an
uncertainty a factor of 1.5 smaller than earlier estimates using polarized DIS
data. This improvement makes the comparison between extracted from
polarized DIS and other techniques a valuable test of QCD.Comment: Published in Phys. Rev. D. V1: 8 pages, 3 figures. V2: Updated
references; Included threshold matching in \alpha_s evolution. Corrected a
typo on the uncertainty for \Lambda_QCD. V3: Published versio
Effect of differences in proton and neutron density distributions on fission barriers
The neutron and proton density distributions obtained in constrained
Hartree-Fock-Bogolyubov calculations with the Gogny force along the fission
paths of 232Th, 236U, 238U and 240Pu are analyzed.
Significant differences in the multipole deformations of neutron and proton
densities are found. The effect on potential energy surfaces and on barrier
heights of an additional constraint imposing similar spatial distributions to
neutrons and protons, as assumed in macroscopic-microscopic models, is studied.Comment: 5 pages in Latex, 4 figures in ep
- …