78,473 research outputs found

    Quark/Gluon Discrimination and Top Tagging with Dual Attention Transformer

    Full text link
    Jet tagging is a crucial classification task in high energy physics. Recently the performance of jet tagging has been significantly improved by the application of deep learning techniques. In this work, we propose Particle Dual Attention Transformer for jet tagging, a new transformer architecture which captures both global information and local information simultaneously. Based on the point cloud representation, we introduce the Channel Attention module to the point cloud transformer and incorporates both the pairwise particle interactions and the pairwise jet feature interactions in the attention mechanism. We demonstrate the effectiveness of the P-DAT architecture in classic top tagging and quark-gluon discrimination tasks, achieving competitive performance compared to other benchmark strategies.Comment: 15 pages, 4 figures, 3 table

    Quantum-inspired Machine Learning on high-energy physics data

    Get PDF
    Tensor Networks, a numerical tool originally designed for simulating quantum many-body systems, have recently been applied to solve Machine Learning problems. Exploiting a tree tensor network, we apply a quantum-inspired machine learning technique to a very important and challenging big data problem in high energy physics: the analysis and classification of data produced by the Large Hadron Collider at CERN. In particular, we present how to effectively classify so-called b-jets, jets originating from b-quarks from proton-proton collisions in the LHCb experiment, and how to interpret the classification results. We exploit the Tensor Network approach to select important features and adapt the network geometry based on information acquired in the learning process. Finally, we show how to adapt the tree tensor network to achieve optimal precision or fast response in time without the need of repeating the learning process. These results pave the way to the implementation of high-frequency real-time applications, a key ingredient needed among others for current and future LHCb event classification able to trigger events at the tens of MHz scale.Comment: 13 pages, 4 figure

    Identifying Heavy-Flavor Jets Using Vectors of Locally Aggregated Descriptors

    Full text link
    Jets of collimated particles serve a multitude of purposes in high energy collisions. Recently, studies of jet interaction with the quark-gluon plasma (QGP) created in high energy heavy ion collisions are of growing interest, particularly towards understanding partonic energy loss in the QGP medium and its related modifications of the jet shower and fragmentation. Since the QGP is a colored medium, the extent of jet quenching and consequently, the transport properties of the medium are expected to be sensitive to fundamental properties of the jets such as the flavor of the parton that initiates the jet. Identifying the jet flavor enables an extraction of the mass dependence in jet-QGP interactions. We present a novel approach to tagging heavy-flavor jets at collider experiments utilizing the information contained within jet constituents via the \texttt{JetVLAD} model architecture. We show the performance of this model in proton-proton collisions at center of mass energy s=200\sqrt{s} = 200 GeV as characterized by common metrics and showcase its ability to extract high purity heavy-flavor jet sample at various jet momenta and realistic production cross-sections including a brief discussion on the impact of out-of-time pile-up. Such studies open new opportunities for future high purity heavy-flavor measurements at jet energies accessible at current and future collider experiments.Comment: 18 pages, 6 figures and 3 tables. Accepted by JINS

    Dynamic texture recognition using time-causal and time-recursive spatio-temporal receptive fields

    Full text link
    This work presents a first evaluation of using spatio-temporal receptive fields from a recently proposed time-causal spatio-temporal scale-space framework as primitives for video analysis. We propose a new family of video descriptors based on regional statistics of spatio-temporal receptive field responses and evaluate this approach on the problem of dynamic texture recognition. Our approach generalises a previously used method, based on joint histograms of receptive field responses, from the spatial to the spatio-temporal domain and from object recognition to dynamic texture recognition. The time-recursive formulation enables computationally efficient time-causal recognition. The experimental evaluation demonstrates competitive performance compared to state-of-the-art. Especially, it is shown that binary versions of our dynamic texture descriptors achieve improved performance compared to a large range of similar methods using different primitives either handcrafted or learned from data. Further, our qualitative and quantitative investigation into parameter choices and the use of different sets of receptive fields highlights the robustness and flexibility of our approach. Together, these results support the descriptive power of this family of time-causal spatio-temporal receptive fields, validate our approach for dynamic texture recognition and point towards the possibility of designing a range of video analysis methods based on these new time-causal spatio-temporal primitives.Comment: 29 pages, 16 figure

    Pulling Out All the Tops with Computer Vision and Deep Learning

    Full text link
    We apply computer vision with deep learning -- in the form of a convolutional neural network (CNN) -- to build a highly effective boosted top tagger. Previous work (the "DeepTop" tagger of Kasieczka et al) has shown that a CNN-based top tagger can achieve comparable performance to state-of-the-art conventional top taggers based on high-level inputs. Here, we introduce a number of improvements to the DeepTop tagger, including architecture, training, image preprocessing, sample size and color pixels. Our final CNN top tagger outperforms BDTs based on high-level inputs by a factor of 2\sim 2--3 or more in background rejection, over a wide range of tagging efficiencies and fiducial jet selections. As reference points, we achieve a QCD background rejection factor of 500 (60) at 50\% top tagging efficiency for fully-merged (non-merged) top jets with pTp_T in the 800--900 GeV (350--450 GeV) range. Our CNN can also be straightforwardly extended to the classification of other types of jets, and the lessons learned here may be useful to others designing their own deep NNs for LHC applications.Comment: 33 pages, 11 figure
    corecore