23,119 research outputs found

    Next-to-leading-order QCD corrections to e+eH+γe^+e^-\to H+\gamma

    Full text link
    The associated production of Higgs boson with a hard photon at lepton collider, i.e., e+eHγe^+e^-\to H\gamma, is known to bear a rather small cross section in Standard Model, and can serve as a sensitive probe for the potential new physics signals. Similar to the loop-induced Higgs decay channels Hγγ,ZγH\to \gamma\gamma, Z\gamma, the e+eHγe^+e^-\to H\gamma process also starts at one-loop order provided that the tiny electron mass is neglected. In this work, we calculate the next-to-leading-order (NLO) QCD corrections to this associated H+γH+\gamma production process, which mainly stem from the gluonic dressing to the top quark loop. The QCD corrections are found to be rather modest at lower center-of-mass energy range (s<300\sqrt{s}<300 GeV), thus of negligible impact on Higgs factory such as CEPC. Nevertheless, when the energy is boosted to the ILC energy range (s400\sqrt{s}\approx 400 GeV), QCD corrections may enhance the leading-order cross section by 20%20\%. In any event, the e+eHγe^+e^-\to H\gamma process has a maximal production rate σmax0.08\sigma_{\rm max}\approx 0.08 fb around s=250\sqrt{s}= 250 GeV, thus CEPC turns out to be the best place to look for this rare Higgs production process. In the high energy limit, the effect of NLO QCD corrections become completely negligible, which can be simply attributed to the different asymptotic scaling behaviors of the LO and NLO cross sections, where the former exhibits a milder decrement 1/s\propto 1/s , but the latter undergoes a much faster decrease 1/s2\propto 1/s^2.Comment: v4, 11 pages, 6 figures, 2 tables; errors in Appendix are fixed; version accepted for publication at PL

    BoostFM: Boosted Factorization Machines for Top-N Feature-based Recommendation

    Get PDF
    Feature-based matrix factorization techniques such as Factorization Machines (FM) have been proven to achieve impressive accuracy for the rating prediction task. However, most common recommendation scenarios are formulated as a top-N item ranking problem with implicit feedback (e.g., clicks, purchases)rather than explicit ratings. To address this problem, with both implicit feedback and feature information, we propose a feature-based collaborative boosting recommender called BoostFM, which integrates boosting into factorization models during the process of item ranking. Specifically, BoostFM is an adaptive boosting framework that linearly combines multiple homogeneous component recommenders, which are repeatedly constructed on the basis of the individual FM model by a re-weighting scheme. Two ways are proposed to efficiently train the component recommenders from the perspectives of both pairwise and listwise Learning-to-Rank (L2R). The properties of our proposed method are empirically studied on three real-world datasets. The experimental results show that BoostFM outperforms a number of state-of-the-art approaches for top-N recommendation

    LambdaFM: Learning Optimal Ranking with Factorization Machines Using Lambda Surrogates

    Get PDF
    State-of-the-art item recommendation algorithms, which apply Factorization Machines (FM) as a scoring function and pairwise ranking loss as a trainer (PRFM for short), have been recently investigated for the implicit feedback based context-aware recommendation problem (IFCAR). However, good recommenders particularly emphasize on the accuracy near the top of the ranked list, and typical pairwise loss functions might not match well with such a requirement. In this paper, we demonstrate, both theoretically and empirically, PRFM models usually lead to non-optimal item recommendation results due to such a mismatch. Inspired by the success of LambdaRank, we introduce Lambda Factorization Machines (LambdaFM), which is particularly intended for optimizing ranking performance for IFCAR. We also point out that the original lambda function suffers from the issue of expensive computational complexity in such settings due to a large amount of unobserved feedback. Hence, instead of directly adopting the original lambda strategy, we create three effective lambda surrogates by conducting a theoretical analysis for lambda from the top-N optimization perspective. Further, we prove that the proposed lambda surrogates are generic and applicable to a large set of pairwise ranking loss functions. Experimental results demonstrate LambdaFM significantly outperforms state-of-the-art algorithms on three real-world datasets in terms of four standard ranking measures

    The Decomposition of Neutron-Antineutron Oscillation Operators

    Full text link
    We study the systematic decomposition of the dimension nine neutron-antineutron oscillation operators at tree and one-loop levels. We discuss the topologies' generation and the assignment of the chiral quarks. The completed lists of the decompositions are provided. We furthermore show an example that the neutron-antineutron oscillation occurs at one-loop level, with the tiny neutrino mass being generated via the scotogenic model and proton decay being evaded.Comment: 27 pages, 6 figures, 19 table
    corecore