1,795 research outputs found

    Kaon decays and the flavour problem

    Get PDF
    After a brief introduction to the so-called flavour problem, we discuss the role of rare K decays in probing the mechanism of quark-flavour mixing. Particular attention is devoted to the formulation of the Minimal Flavour Violation hypothesis, as a general and natural solution to the flavour problem, and to the fundamental role of K -> pi nu nu-bar decays in testing this scenario.Comment: 10 pages, 6 figures, contribution to TH 2002 (Paris, July 2002

    K^+ -> pi^+pi^0e^+e^-: a novel short-distance probe

    Full text link
    We study the decay K^+ -> pi^+ pi^0 e^+ e^-, currently under analysis by the NA62 Collaboration at CERN. In particular, we provide a detailed analysis of the Dalitz plot for the long-distance, gamma^*-mediated, contributions (Bremsstrahlung, direct emission and its interference). We also examine a set of asymmetries to isolate genuine short-distance effects. While we show that charge asymmetries are not required to test short distances, they provide the best environment for its detection. This constitutes by itself a strong motivation for NA62 to study K^- decays in the future. We therefore provide a detailed study of different charge asymmetries and the corresponding estimated signals. Whenever possible, we make contact with the related processes K^+ -> pi^+ pi^0 gamma and K_L -> pi^+ pi^- e^+ e^- and discuss the advantages of K^+ -> pi^+ pi^0 e^+ e^- over them.Comment: 25 pages, 6 figure

    Long-distance contribution to the forward-backward asymmetry in decays K+ --> pi+ l+ l-

    Full text link
    The long-distance contribution via the two-photon intermediate state to the forward-backward asymmetries in decays K+ --> pi+ l+ l- (l=e and mu) has been studied within the standard model. In order to evaluate the dispersive part of the K+ --> pi+ gamma* gamma* --> pi+ l+ l- amplitude, we employ a phenomenological form factor to soften the ultraviolet behavior of the transition. It is found that, this long-distance transition, although subject to some theoretical uncertainties, can lead to significant contributions to the forward-backward asymmetries, which could be tested in the future high-precise experiments.Comment: 13 pages, 5 figure

    FROM THE DIGITAL TWIN OF ARCHITECTURAL HERITAGE TO THE DEFINITION OF THE CONSTRAINT MAPS. THE CASE STUDY OF THE CASTELLO SFORZESCO IN MILAN

    Get PDF
    The work presented here is part of the widespread digitisation process that is affecting the world of architecture and cultural heritage. The text reports part of the activities carried out on the Castello Sforzesco in Milan, in collaboration with its Superintendence bodies. This work proposes an innovative way to draft, in a digital environment, 3D constraint maps, capable of describing and protecting the complex system of structures in that area, especially in the subsoil. The activity began with the creation of a georeferenced 3D model of all the existing architectural elements in the analyzed area of the Castle. The work was carried out on the northwest corner of the fortress, the area between the Torre della Colubrina and the Porta del Soccorso. The modeling activity was based both on data obtained from surveys and historical documentation describing structures that are no longer visible. From the 3D model, it was then possible to obtain all useful data to draw up constraint maps in GIS (Geographic Information System) environment. This latter allows multiple interactions both with the normal surface cartographic information (Topographic Database and others) and with the multitude of datasets available for managing the territory and its subsystems. The constraints consider a ‘safety’ offset, applied in all three directions. These maps describe the various three-dimensional areas, referring to the different structures present in the subsoil, within which it is recommended to operate with extreme caution, in relation to the position and conformation of the same identified assets

    Geometric Hardy inequalities for the sub-elliptic Laplacian on convex domains in the Heisenberg group

    Full text link
    We prove geometric LpL^p versions of Hardy's inequality for the sub-elliptic Laplacian on convex domains Ω\Omega in the Heisenberg group Hn\mathbb{H}^n, where convex is meant in the Euclidean sense. When p=2p=2 and Ω\Omega is the half-space given by ⟚Ο,Μ⟩>d\langle \xi, \nu\rangle > d this generalizes an inequality previously obtained by Luan and Yang. For such pp and Ω\Omega the inequality is sharp and takes the form \begin{equation} \int_\Omega |\nabla_{\mathbb{H}^n}u|^2 \, d\xi \geq \frac{1}{4}\int_{\Omega} \sum_{i=1}^n\frac{\langle X_i(\xi), \nu\rangle^2+\langle Y_i(\xi), \nu\rangle^2}{\textrm{dist}(\xi, \partial \Omega)^2}|u|^2\, d\xi, \end{equation} where dist( ⋅ ,∂Ω)\textrm{dist}(\, \cdot\,, \partial \Omega) denotes the Euclidean distance from ∂Ω\partial \Omega.Comment: 14 page

    A holographic approach to low-energy weak interactions of hadrons

    Full text link
    We apply the double-trace formalism to incorporate nonleptonic weak interactions of hadrons into holographic models of the strong interactions. We focus our attention upon ΔS=1\Delta S=1 nonleptonic kaon decays. By working with a Yang-Mills--Chern-Simons 5-dimensional action, we explicitly show how, at low energies, one recovers the ΔS=1\Delta S=1 weak chiral Lagrangian for both the anomalous and nonanomalous sectors. We provide definite predictions for the low energy coefficients in terms of the AdS metric and argue that the double-trace formalism is a 5-dimensional avatar of the Weak Deformation Model introduced long ago by Ecker et al. As a significant phenomenological application, we reassess the K→3πK\to 3\pi decays in the light of the holographic model. Previous models found a fine-tuned cancellation of resonance exchange in these decays, which was both conceptually puzzling and quantitatively in disagreement with experimental results. The holographic model we build is an illustrative counterexample showing that the cancellation encountered in the literature is not generic but a model-dependent statement and that agreement with experiment can be obtained.Comment: 20 page

    TORCH: A Cherenkov Based Time-of-Flight Detector

    Get PDF
    TORCH is a novel high-precision time-of-flight detector suitable for large area applications and covering the momentum range up to 10 GeV/c. The concept uses Cherenkov photons produced in a fused silica radiator which are propagated to focussing optics coupled to fast photodetectors. For this purpose, custom MCP-PMTs are being produced in collaboration with industrial partners. The development is divided into three phases. Phase 1 addresses the lifetime requirements for TORCH, Phase 2 will customize the MCP-PMT granularity and Phase 3 will deliver prototypes that meet the TORCH requirements. Phase 1 devices have been successfully delivered and initial tests show stable gain performance for integrated anode current >5 C/cm2 and a single photon time resolution of ≀ 30 ps. Initial simulations indicate the single photon timing resolution of the TORCH detector will be ∌70 ps

    The s ---> d gamma decay in and beyond the Standard Model

    Get PDF
    The New Physics sensitivity of the s ---> d gamma transition and its accessibility through hadronic processes are thoroughly investigated. Firstly, the Standard Model predictions for the direct CP-violating observables in radiative K decays are systematically improved. Besides, the magnetic contribution to epsilon prime is estimated and found subleading, even in the presence of New Physics, and a new strategy to resolve its electroweak versus QCD penguin fraction is identified. Secondly, the signatures of a series of New Physics scenarios, characterized as model-independently as possible in terms of their underlying dynamics, are investigated by combining the information from all the FCNC transitions in the s ---> d sector.Comment: 54 pages, 14 eps figure

    Iso-singlet Down Quark Mixing And CP Violation Experiments

    Full text link
    We confront the new physics models with extra iso-singlet down quarks in the new CP violation experimental era with sin⁥(2ÎČ)\sin{(2\beta)} and Ï”â€Č/Ï”\epsilon'/\epsilon measurements, K+→π+ΜΜˉK^+ \to \pi^+ \nu \bar{\nu} events, and xsx_s limits. The closeness of the new experimental results to the standard model theory requires us to include full SM amplitudes in the analysis. In models allowing mixing to a new isosinglet down quark, as in E6_6, flavor changing neutral currents are induced that allow a Z0Z^0 mediated contribution to B−BˉB-\bar B mixing and which bring in new phases. In (ρ,η)(\rho,\eta), (xs,sin⁥(Îł))(x_s,\sin{(\gamma)}), and (xs,sin⁥(2ϕs))(x_s, \sin{(2\phi_s)}) plots we still find much larger regions in the four down quark model than in the SM, reaching down to η≈0\eta \approx 0, 0≀sin⁥(Îł)≀10 \leq \sin{(\gamma)} \leq 1, −.75≀sin⁥(2α)≀0.15-.75 \leq \sin{(2\alpha)} \leq 0.15, and sin⁥(2ϕs)\sin{(2\phi_s)} down to zero, all at 1σ\sigma. We elucidate the nature of the cancellation in an order λ5\lambda^5 four down quark mixing matrix element which satisfies the experiments and reduces the number of independent angles and phases. We also evaluate tests of unitarity for the 3×33\times3 CKM submatrix.Comment: 14 pages, 16 figures, REVTeX
    • 

    corecore