1,510 research outputs found

    Higgs properties measurements using the four lepton decay channel

    Full text link
    The measurements of the properties of the Higgs boson are presented in the H\rightarrowZZ\rightarrow4\ell (\ell=e,μ\mu) decay channel using a data sample corresponding to an integrated luminosity of 35.9 fb1^{-1} of proton-proton collisions at a center-of-mass energy of 13 TeV recorded by the CMS detector at the LHC. The signal-strength modifier μ\mu, defined as the production cross section of the Higgs boson times its branching fraction to four leptons relative to the standard model expectation, is measured to be μ=1.050.17+0.19\mu=1.05^{+0.19}_{-0.17} at mH=125.09 GeVm_{\mathrm{H}}=125.09~\mathrm{GeV}. Constraints are set on the strength modifiers for the main Higgs boson production modes. The mass is measured to be mH=125.26±0.21 GeVm_{\mathrm{H}}=125.26 \pm 0.21~\mathrm{GeV} and the width is constrained using on-shell production to be ΓH<1.10 GeV\Gamma_{\mathrm{H}}<1.10~\mathrm{GeV}, at 95%95\% CL. The fiducial cross section is measured to be 2.900.44+0.48(stat.)0.22+0.27(sys.) fb2.90^{+0.48}_{-0.44}({\rm stat.})^{+0.27}_{-0.22}({\rm sys.})~{\mathrm{fb}}, which is compatible with the standard model prediction of 2.72±0.14 fb2.72\pm0.14~{\mathrm{fb}}.Comment: Presented at LHCP201

    A discussion of H → ZZ → 4l results and analysis strategies for 13TeV

    Get PDF
    Studies of Higgs boson production are presented using the H → ZZ → 4l (l =e,μ) decay. These studies are performed using a data sample corresponding to an integrated luminosity of 2.8fb−1 of pp collisions at a center-of-mass energy of 13TeV collected by the CMS experiment at the LHC during 2015. The observed significance for the standard model Higgs boson with mH = 125 .09GeV is 2.5σ, where the expected significance is 3.4σ. The model-independent fiducial cross section is measured to be σfid =2 .48+1.48 −1.14(stat⊕sys)+0.01 −0.04(model dep.)f

    Fluorescent-based nanosensors for selective detection of a wide range of biological macromolecules: A comprehensive review

    Get PDF
    Thanks to their unique attributes, such as good sensitivity, selectivity, high surface-to-volume ratio, and versatile optical and electronic properties, fluorescent-based bioprobes have been used to create highly sensitive nano -biosensors to detect various biological and chemical agents. These sensors are superior to other analytical instrumentation techniques like gas chromatography, high-performance liquid chromatography, and capillary electrophoresis for being biodegradable, eco-friendly, and more economical, operational, and cost-effective. Moreover, several reports have also highlighted their application in the early detection of biomarkers associ-ated with drug-induced organ damage such as liver, kidney, or lungs. In the present work, we comprehensively overviewed the electrochemical sensors that employ nanomaterials (nanoparticles/colloids or quantum dots, carbon dots, or nanoscaled metal-organic frameworks, etc.) to detect a variety of biological macromolecules based on fluorescent emission spectra. In addition, the most important mechanisms and methods to sense amino acids, protein, peptides, enzymes, carbohydrates, neurotransmitters, nucleic acids, vitamins, ions, metals, and electrolytes, blood gases, drugs (i.e., anti-inflammatory agents and antibiotics), toxins, alkaloids, antioxidants, cancer biomarkers, urinary metabolites (i.e., urea, uric acid, and creatinine), and pathogenic microorganisms were outlined and compared in terms of their selectivity and sensitivity. Altogether, the small dimensions and capability of these nanosensors for sensitive, label-free, real-time sensing of chemical, biological, and pharma-ceutical agents could be used in array-based screening and in-vitro or in-vivo diagnostics. Although fluorescent nanoprobes are widely applied in determining biological macromolecules, unfortunately, they present many challenges and limitations. Efforts must be made to minimize such limitations in utilizing such nanobiosensors with an emphasis on their commercial developments. We believe that the current review can foster the wider incorporation of nanomedicine and will be of particular interest to researchers working on fluorescence tech-nology, material chemistry, coordination polymers, and related research areas

    IN-DEPTH REVIEW OF GREEN HUMAN RESOURCE MANAGEMENT ALONG WITH LATEST INITIATIVES

    Get PDF
    Purpose: The paper aims to review advancements and current literature in the field of Green Human Resource Management (GHRM). It further intends to understand the relationship of GHRM with multiple variables and to highlight multiple green initiatives that organizations can use to further their environmental agenda

    Optimasi Portofolio Resiko Menggunakan Model Markowitz MVO Dikaitkan dengan Keterbatasan Manusia dalam Memprediksi Masa Depan dalam Perspektif Al-Qur`an

    Full text link
    Risk portfolio on modern finance has become increasingly technical, requiring the use of sophisticated mathematical tools in both research and practice. Since companies cannot insure themselves completely against risk, as human incompetence in predicting the future precisely that written in Al-Quran surah Luqman verse 34, they have to manage it to yield an optimal portfolio. The objective here is to minimize the variance among all portfolios, or alternatively, to maximize expected return among all portfolios that has at least a certain expected return. Furthermore, this study focuses on optimizing risk portfolio so called Markowitz MVO (Mean-Variance Optimization). Some theoretical frameworks for analysis are arithmetic mean, geometric mean, variance, covariance, linear programming, and quadratic programming. Moreover, finding a minimum variance portfolio produces a convex quadratic programming, that is minimizing the objective function ðð¥with constraintsð ð 𥠥 ðandð´ð¥ = ð. The outcome of this research is the solution of optimal risk portofolio in some investments that could be finished smoothly using MATLAB R2007b software together with its graphic analysis

    Search for supersymmetry in events with one lepton and multiple jets in proton-proton collisions at root s=13 TeV

    Get PDF
    Peer reviewe

    Measurement of the top quark forward-backward production asymmetry and the anomalous chromoelectric and chromomagnetic moments in pp collisions at √s = 13 TeV

    Get PDF
    Abstract The parton-level top quark (t) forward-backward asymmetry and the anomalous chromoelectric (d̂ t) and chromomagnetic (μ̂ t) moments have been measured using LHC pp collisions at a center-of-mass energy of 13 TeV, collected in the CMS detector in a data sample corresponding to an integrated luminosity of 35.9 fb−1. The linearized variable AFB(1) is used to approximate the asymmetry. Candidate t t ¯ events decaying to a muon or electron and jets in final states with low and high Lorentz boosts are selected and reconstructed using a fit of the kinematic distributions of the decay products to those expected for t t ¯ final states. The values found for the parameters are AFB(1)=0.048−0.087+0.095(stat)−0.029+0.020(syst),μ̂t=−0.024−0.009+0.013(stat)−0.011+0.016(syst), and a limit is placed on the magnitude of | d̂ t| &lt; 0.03 at 95% confidence level. [Figure not available: see fulltext.

    Measurement of the top quark mass using charged particles in pp collisions at root s=8 TeV

    Get PDF
    Peer reviewe

    Measurement of t(t)over-bar normalised multi-differential cross sections in pp collisions at root s=13 TeV, and simultaneous determination of the strong coupling strength, top quark pole mass, and parton distribution functions

    Get PDF
    Peer reviewe

    An embedding technique to determine ττ backgrounds in proton-proton collision data

    Get PDF
    An embedding technique is presented to estimate standard model tau tau backgrounds from data with minimal simulation input. In the data, the muons are removed from reconstructed mu mu events and replaced with simulated tau leptons with the same kinematic properties. In this way, a set of hybrid events is obtained that does not rely on simulation except for the decay of the tau leptons. The challenges in describing the underlying event or the production of associated jets in the simulation are avoided. The technique described in this paper was developed for CMS. Its validation and the inherent uncertainties are also discussed. The demonstration of the performance of the technique is based on a sample of proton-proton collisions collected by CMS in 2017 at root s = 13 TeV corresponding to an integrated luminosity of 41.5 fb(-1).Peer reviewe
    corecore