244 research outputs found

    Learning the language of QCD jets with transformers

    Full text link
    Transformers have become the primary architecture for natural language processing. In this study, we explore their use for auto-regressive density estimation in high-energy jet physics, which involves working with a high-dimensional space. We draw an analogy between sentences and words in natural language and jets and their constituents in high-energy physics. Specifically, we investigate density estimation for light QCD jets and hadronically decaying boosted top jets. Since transformers allow easy sampling from learned densities, we exploit their generative capability to assess the quality of the density estimate. Our results indicate that the generated data samples closely resemble the original data, as evidenced by the excellent agreement of distributions such as particle multiplicity or jet mass. Furthermore, the generated samples are difficult to distinguish from the original data, even by a powerful supervised classifier. Given their exceptional data processing capabilities, transformers could potentially be trained directly on the massive LHC data sets to learn the probability densities in high-energy jet physics.Comment: Few references added; Version accepted for publication by JHE

    Framework for automatic production simulation tuning with machine learning

    Get PDF
    Production system simulation is a powerful tool for optimizing the use of resources on both the planning and control level. However, creating and tuning such models manually is a tedious and error-prone task. Despite some approaches to automate this process, the state-of-the-art relies on the generation of models, by incorporating the knowledge of experts. Nevertheless, effectively creating and tuning such production simulations is, thus, a key driver for reducing costs, carbon footprint, and tardiness and therefore an essential factor in today´s production. Beneficial would be automated and flexible frameworks, since these are applicable to different use cases requiring less effort. Yet, in the age of Industry 4.0, data is ubiquitous and easily available and can serve as a basis for virtual models representing reality. Increasingly, these virtual models shall be interlinked with the current state of real-world systems to form so-called digital twins. As automated and flexible frameworks are missing, this paper proposes a novel approach where observed real system behavior is used and fed into a large-scale machine learning model trained on a plethora of possible parameter sets. The main target is to train this machine learning model to minimize the reality gap between the behavior of the simulated and real system by selecting corresponding simulation system parameters. By estimating those parameters an enhancement of the simulation will emerge. An interlink to real systems can be derived resulting in a digital shadow which is capable to forecast the future similarly to reality. The approach to overcoming the gap between reality and simulation (real2sim) is validated in simulations

    Design and Implementation of a Distributed Ledger Technology Platform to Support Customs Processes within Supply Chains

    Get PDF
    In international trade, customs clearance fulfills complex and country-specific tasks in the execution of supply chain processes. Importers and exporters have to integrate customs authorities into the information flow, as customs authorities require information, e.g., on the bill of lading and the commercial invoice apart from the customs declaration. In addition, involved sub-service providers increase the problem of information asymmetry and the required coordination effort. Practice and research consider Distributed Ledger Technology (DLT) as a potential solution since this technology maintains a mutually agreed and secure database of value-creation partners. However, research has hardly investigated the design of such DLT systems. Therefore, we present a requirements catalogue, a concept, and a prototype of a DLT platform to address the outlined problem of information asymmetry, especially with a focus on customs processes

    Back To The Roots: Tree-Based Algorithms for Weakly Supervised Anomaly Detection

    Full text link
    Weakly supervised methods have emerged as a powerful tool for model-agnostic anomaly detection at the Large Hadron Collider (LHC). While these methods have shown remarkable performance on specific signatures such as di-jet resonances, their application in a more model-agnostic manner requires dealing with a larger number of potentially noisy input features. In this paper, we show that using boosted decision trees as classifiers in weakly supervised anomaly detection gives superior performance compared to deep neural networks. Boosted decision trees are well known for their effectiveness in tabular data analysis. Our results show that they not only offer significantly faster training and evaluation times, but they are also robust to a large number of noisy input features. By using advanced gradient boosted decision trees in combination with ensembling techniques and an extended set of features, we significantly improve the performance of weakly supervised methods for anomaly detection at the LHC. This advance is a crucial step towards a more model-agnostic search for new physics.Comment: 11 pages, 9 figure

    Quantitative Multi-Parameter Mapping Optimized for the Clinical Routine

    Get PDF
    Using quantitative multi-parameter mapping (MPM), studies can investigate clinically relevant microstructural changes with high reliability over time and across subjects and sites. However, long acquisition times (20 min for the standard 1-mm isotropic protocol) limit its translational potential. This study aimed to evaluate the sensitivity gain of a fast 1.6-mm isotropic MPM protocol including post-processing optimized for longitudinal clinical studies. 6 healthy volunteers (35 +/- 7 years old; 3 female) were scanned at 3T to acquire the following whole-brain MPM maps with 1.6 mm isotropic resolution: proton density (PD), magnetization transfer saturation (MT), longitudinal relaxation rate (R1), and transverse relaxation rate (R2*). MPM maps were generated using two RF transmit field (B1+) correction methods: (1) using an acquired B1+ map and (2) using a data-driven approach. Maps were generated with and without Gibb's ringing correction. The intra-/inter-subject coefficient of variation (CoV) of all maps in the gray and white matter, as well as in all anatomical regions of a fine-grained brain atlas, were compared between the different post-processing methods using Student's t-test. The intra-subject stability of the 1.6-mm MPM protocol is 2-3 times higher than for the standard 1-mm sequence and can be achieved in less than half the scan duration. Intra-subject variability for all four maps in white matter ranged from 1.2-5.3% and in gray matter from 1.8 to 9.2%. Bias-field correction using an acquired B1+ map significantly improved intra-subject variability of PD and R1 in the gray (42%) and white matter (54%) and correcting the raw images for the effect of Gibb's ringing further improved intra-subject variability in all maps in the gray (11%) and white matter (10%). Combining Gibb's ringing correction and bias field correction using acquired B1+ maps provides excellent stability of the 7-min MPM sequence with 1.6 mm resolution suitable for the clinical routine

    Standardization of T1w/T2w Ratio Improves Detection of Tissue Damage in Multiple Sclerosis

    Get PDF
    Normal appearing white matter (NAWM) damage develops early in multiple sclerosis (MS) and continues in the absence of new lesions. The ratio of T1w and T2w (T1w/T2w ratio), a measure of white matter integrity, has previously shown reduced intensity values in MS NAWM. We evaluate the validity of a standardized T1w/T2w ratio (sT1w/T2w ratio) in MS and whether this method is sensitive in detecting MS-related differences in NAWM. T1w and T2w scans were acquired at 3 Tesla in 47 patients with relapsing-remitting MS and 47 matched controls (HC). T1w/T2w and sT1w/T2w ratios were then calculated. We compared between-group variability between T1w/T2w and sT1w/T2w ratio in HC and MS and assessed for group differences. We also evaluated the relationship between the T1w/T2w and sT1w/T2w ratios and clinically relevant variables. Compared to the classic T1w/T2w ratio, the between-subject variability in sT1w/T2w ratio showed a significant reduction in MS patients (0 <. 0.001) and HC < 0.001). However, only sT1w/T2w ratio values were reduced in patients compared to HC (p < 0.001). The sT1w/T2w ratio intensity values were significantly influenced by age, T2 lesion volume and group status (MS vs. HC) (adjusted R-2 = 0.30, p 0.001). We demonstrate the validity of the sT1w/T2w ratio in MS and that it is more sensitive to MS-related differences in NAWM compared to T1w/T2w ratio. The sT1w/T2w ratio shows promise as an easily-implemented measure of NAWM in MS using readily available scans and simple post-processing methods

    Characterizing the outer ear transfer function in dependence of interindividual differences of outer ear geometry

    Get PDF
    The outer ear transfer function can be used to describe the influence of the outer ear canal and its geometric variance in cross-section as well as its path on the sound field in the ear canal and the sound pressure level resulting at the ear drum. The variance of outer ear geometry is described by analysis of polysiloxane castings of the outer ear. Algorithms are developed to determine various parameters of the outer ear geometry and to gain access on a huge amount of data (over 100.000 data sets). Sound transmission in form of the outer ear transfer function is analyzed for various outer ear geometries using a finite element model as well as an experimental setup. In both cases sound (frequency band: 20 Hz to 20 kHz) is send to a model of the outer ear as a plane wave parallel to the plane of the Pinna
    • …
    corecore