486 research outputs found
Sample entropy analysis of EEG signals via artificial neural networks to model patients' consciousness level based on anesthesiologists experience.
Electroencephalogram (EEG) signals, as it can express the human brain's activities and reflect awareness, have been widely used in many research and medical equipment to build a noninvasive monitoring index to the depth of anesthesia (DOA). Bispectral (BIS) index monitor is one of the famous and important indicators for anesthesiologists primarily using EEG signals when assessing the DOA. In this study, an attempt is made to build a new indicator using EEG signals to provide a more valuable reference to the DOA for clinical researchers. The EEG signals are collected from patients under anesthetic surgery which are filtered using multivariate empirical mode decomposition (MEMD) method and analyzed using sample entropy (SampEn) analysis. The calculated signals from SampEn are utilized to train an artificial neural network (ANN) model through using expert assessment of consciousness level (EACL) which is assessed by experienced anesthesiologists as the target to train, validate, and test the ANN. The results that are achieved using the proposed system are compared to BIS index. The proposed system results show that it is not only having similar characteristic to BIS index but also more close to experienced anesthesiologists which illustrates the consciousness level and reflects the DOA successfully.This research is supported by the Center forDynamical Biomarkers and Translational Medicine, National Central University, Taiwan, which is sponsored by Ministry of Science and Technology (Grant no. MOST103-2911-I-008-001). Also, it is supported by National Chung-Shan Institute of Science & Technology in Taiwan (Grant nos. CSIST-095-V301 and CSIST-095-V302)
Can we apply the Mendelian randomization methodology without considering epigenetic effects?
<p>Abstract</p> <p>Introduction</p> <p>Instrumental variable (IV) methods have been used in econometrics for several decades now, but have only recently been introduced into the epidemiologic research frameworks. Similarly, Mendelian randomization studies, which use the IV methodology for analysis and inference in epidemiology, were introduced into the epidemiologist's toolbox only in the last decade.</p> <p>Analysis</p> <p>Mendelian randomization studies using instrumental variables (IVs) have the potential to avoid some of the limitations of observational epidemiology (confounding, reverse causality, regression dilution bias) for making causal inferences. Certain limitations of randomized controlled trials, such as problems with generalizability, feasibility and ethics for some exposures, and high costs, also make the use of Mendelian randomization in observational studies attractive. Unlike conventional randomized controlled trials (RCTs), Mendelian randomization studies can be conducted in a representative sample without imposing any exclusion criteria or requiring volunteers to be amenable to random treatment allocation.</p> <p>Within the last decade, epigenetics has gained recognition as an independent field of study, and appears to be the new direction for future research into the genetics of complex diseases. Although previous articles have addressed some of the limitations of Mendelian randomization (such as the lack of suitable genetic variants, unreliable associations, population stratification, linkage disequilibrium (LD), pleiotropy, developmental canalization, the need for large sample sizes and some potential problems with binary outcomes), none has directly characterized the impact of epigenetics on Mendelian randomization. The possibility of epigenetic effects (non-Mendelian, heritable changes in gene expression not accompanied by alterations in DNA sequence) could alter the core instrumental variable assumptions of Mendelian randomization.</p> <p>This paper applies conceptual considerations, algebraic derivations and data simulations to question the appropriateness of Mendelian randomization methods when epigenetic modifications are present.</p> <p>Conclusion</p> <p>Given an inheritance of gene expression from parents, Mendelian randomization studies not only need to assume a random distribution of alleles in the offspring, but also a random distribution of epigenetic changes (e.g. gene expression) at conception, in order for the core assumptions of the Mendelian randomization methodology to remain valid. As an increasing number of epidemiologists employ Mendelian randomization methods in their research, caution is therefore needed in drawing conclusions from these studies if these assumptions are not met.</p
Fast splice site detection using information content and feature reduction
Background: Accurate identification of splice sites in DNA sequences plays a key role in the prediction of gene structure in eukaryotes. Already many computational methods have been proposed for the detection of splice sites and some of them showed high prediction accuracy. However, most of these methods are limited in terms of their long computation time when applied to whole genome sequence data. Results: In this paper we propose a hybrid algorithm which combines several effective and informative input features with the state of the art support vector machine (SVM). To obtain the input features we employ information content method based on Shannon\u27s information theory, Shapiro\u27s score scheme, and Markovian probabilities. We also use a feature elimination scheme to reduce the less informative features from the input data. Conclusion: In this study we propose a new feature based splice site detection method that shows improved acceptor and donor splice site detection in DNA sequences when the performance is compared with various state of the art and well known method
Performance of CMS muon reconstruction in pp collision events at sqrt(s) = 7 TeV
The performance of muon reconstruction, identification, and triggering in CMS
has been studied using 40 inverse picobarns of data collected in pp collisions
at sqrt(s) = 7 TeV at the LHC in 2010. A few benchmark sets of selection
criteria covering a wide range of physics analysis needs have been examined.
For all considered selections, the efficiency to reconstruct and identify a
muon with a transverse momentum pT larger than a few GeV is above 95% over the
whole region of pseudorapidity covered by the CMS muon system, abs(eta) < 2.4,
while the probability to misidentify a hadron as a muon is well below 1%. The
efficiency to trigger on single muons with pT above a few GeV is higher than
90% over the full eta range, and typically substantially better. The overall
momentum scale is measured to a precision of 0.2% with muons from Z decays. The
transverse momentum resolution varies from 1% to 6% depending on pseudorapidity
for muons with pT below 100 GeV and, using cosmic rays, it is shown to be
better than 10% in the central region up to pT = 1 TeV. Observed distributions
of all quantities are well reproduced by the Monte Carlo simulation.Comment: Replaced with published version. Added journal reference and DO
Performance of CMS muon reconstruction in pp collision events at sqrt(s) = 7 TeV
The performance of muon reconstruction, identification, and triggering in CMS
has been studied using 40 inverse picobarns of data collected in pp collisions
at sqrt(s) = 7 TeV at the LHC in 2010. A few benchmark sets of selection
criteria covering a wide range of physics analysis needs have been examined.
For all considered selections, the efficiency to reconstruct and identify a
muon with a transverse momentum pT larger than a few GeV is above 95% over the
whole region of pseudorapidity covered by the CMS muon system, abs(eta) < 2.4,
while the probability to misidentify a hadron as a muon is well below 1%. The
efficiency to trigger on single muons with pT above a few GeV is higher than
90% over the full eta range, and typically substantially better. The overall
momentum scale is measured to a precision of 0.2% with muons from Z decays. The
transverse momentum resolution varies from 1% to 6% depending on pseudorapidity
for muons with pT below 100 GeV and, using cosmic rays, it is shown to be
better than 10% in the central region up to pT = 1 TeV. Observed distributions
of all quantities are well reproduced by the Monte Carlo simulation.Comment: Replaced with published version. Added journal reference and DO
X-ray emission from the Sombrero galaxy: discrete sources
We present a study of discrete X-ray sources in and around the
bulge-dominated, massive Sa galaxy, Sombrero (M104), based on new and archival
Chandra observations with a total exposure of ~200 ks. With a detection limit
of L_X = 1E37 erg/s and a field of view covering a galactocentric radius of ~30
kpc (11.5 arcminute), 383 sources are detected. Cross-correlation with Spitler
et al.'s catalogue of Sombrero globular clusters (GCs) identified from HST/ACS
observations reveals 41 X-rays sources in GCs, presumably low-mass X-ray
binaries (LMXBs). We quantify the differential luminosity functions (LFs) for
both the detected GC and field LMXBs, whose power-low indices (~1.1 for the
GC-LF and ~1.6 for field-LF) are consistent with previous studies for
elliptical galaxies. With precise sky positions of the GCs without a detected
X-ray source, we further quantify, through a fluctuation analysis, the GC LF at
fainter luminosities down to 1E35 erg/s. The derived index rules out a
faint-end slope flatter than 1.1 at a 2 sigma significance, contrary to recent
findings in several elliptical galaxies and the bulge of M31. On the other
hand, the 2-6 keV unresolved emission places a tight constraint on the field
LF, implying a flattened index of ~1.0 below 1E37 erg/s. We also detect 101
sources in the halo of Sombrero. The presence of these sources cannot be
interpreted as galactic LMXBs whose spatial distribution empirically follows
the starlight. Their number is also higher than the expected number of cosmic
AGNs (52+/-11 [1 sigma]) whose surface density is constrained by deep X-ray
surveys. We suggest that either the cosmic X-ray background is unusually high
in the direction of Sombrero, or a distinct population of X-ray sources is
present in the halo of Sombrero.Comment: 11 figures, 5 tables, ApJ in pres
Azimuthal anisotropy of charged particles at high transverse momenta in PbPb collisions at sqrt(s[NN]) = 2.76 TeV
The azimuthal anisotropy of charged particles in PbPb collisions at
nucleon-nucleon center-of-mass energy of 2.76 TeV is measured with the CMS
detector at the LHC over an extended transverse momentum (pt) range up to
approximately 60 GeV. The data cover both the low-pt region associated with
hydrodynamic flow phenomena and the high-pt region where the anisotropies may
reflect the path-length dependence of parton energy loss in the created medium.
The anisotropy parameter (v2) of the particles is extracted by correlating
charged tracks with respect to the event-plane reconstructed by using the
energy deposited in forward-angle calorimeters. For the six bins of collision
centrality studied, spanning the range of 0-60% most-central events, the
observed v2 values are found to first increase with pt, reaching a maximum
around pt = 3 GeV, and then to gradually decrease to almost zero, with the
decline persisting up to at least pt = 40 GeV over the full centrality range
measured.Comment: Replaced with published version. Added journal reference and DO
Compressed representation of a partially defined integer function over multiple arguments
In OLAP (OnLine Analitical Processing) data are analysed in an n-dimensional cube. The cube may be represented as a partially defined function over n arguments. Considering that often the function is not defined everywhere, we ask: is there a known way of representing the function or the points in which it is defined, in a more compact manner than the trivial one
Search for new physics with same-sign isolated dilepton events with jets and missing transverse energy
A search for new physics is performed in events with two same-sign isolated
leptons, hadronic jets, and missing transverse energy in the final state. The
analysis is based on a data sample corresponding to an integrated luminosity of
4.98 inverse femtobarns produced in pp collisions at a center-of-mass energy of
7 TeV collected by the CMS experiment at the LHC. This constitutes a factor of
140 increase in integrated luminosity over previously published results. The
observed yields agree with the standard model predictions and thus no evidence
for new physics is found. The observations are used to set upper limits on
possible new physics contributions and to constrain supersymmetric models. To
facilitate the interpretation of the data in a broader range of new physics
scenarios, information on the event selection, detector response, and
efficiencies is provided.Comment: Published in Physical Review Letter
- …