6,931 research outputs found
Tree Level Unitarity Bounds for the Minimal B-L Model
We have derived the unitarity bounds in the high energy limit for the minimal
B-L extension of the Standard Model by analysing the full class of Higgs and
would-be Goldstone boson two-to-two scatterings at tree level. Moreover, we
have investigated how these limits could vary at some lower critical value of
the energy.Comment: 20 pages, 4 figures, 2 tables; 1d figure modified, typos corrected,
bibliography augmented; published in PRD after minor adjustmen
The Z' boson of the minimal B-L model at future Linear Colliders in e+e- --> mu+mu-
We study the capabilities of future electron-positron Linear Colliders, with
centre-of-mass energy at the TeV scale, in accessing the parameter space of a
boson within the minimal model. We carry out a detailed comparison
between the discovery regions mapped over a two-dimensional configuration space
( mass and coupling) at the Large Hadron Collider and possible future
Linear Colliders for the case of di-muon production. As known in the literature
for other models, we confirm that leptonic machines, as compared to the
CERN hadronic accelerator, display an additional potential in discovering a
boson as well as in allowing one to study its properties at a level of
precision well beyond that of any of the existing colliders.Comment: 5 pages, proceeding of LC09 (Perugia), published by the Italian
Physical Society in the Nuovo Cimento C (Colloquia
Dynamic reconfiguration of GCM components
We detail in this report past research and current/future developments in formal specification of Grid component systems by temporal logic and consequent resolution technique, for an automated dynamic reconfiguration of components. It is analysed the specification procedure of GCM (Grid Component Model) components and infrastructure in respect to their state behaviour, and the verification process in a dynamic and reconfigurable distributed system. Furthermore it is demonstrated how an automata based method is used to achieve the specification, as well as how the enrichment of the temporal specification language of Computation Tree Logic CTL with the ability to capture norms, allows to formally define the concept of reconfiguration
Phenomenology of the minimal B-L extension of the Standard Model
We present the Large Hadron Collider (LHC) discovery potential in the
and heavy neutrino sectors of a enlarged Standard Model also
encompassing three heavy Majorana neutrinos. This model exhibits novel
signatures at the LHC, the most interesting arising from a decay chain
involving heavy neutrinos, eventually decaying into leptons and jets. In
particular, this signature allows one to measure the and heavy neutrino
masses involved. In addition, over a large region of parameter space, the heavy
neutrinos are rather long-lived particles producing distinctive displaced
vertices that can be seen in the detectors. Lastly, the simultaneous
measurement of both the heavy neutrino mass and decay length enables an
estimate of the absolute mass of the parent light neutrino. For completeness,
we will also compare the LHC and a future Linear Collider (LC) discovery
potentials.Comment: 4 pages, no figures. LaTeX. Talk given at "The 2009 Europhysics
Conference on High Energy Physics", Krakow, Poland, July 16-22, 200
INTEGRATED PHOTOGRAMMETRIC SURVEY AND BIM MODELLING FOR THE PROTECTION OF SCHOOL HERITAGE, APPLICATIONS ON A CASE STUDY
The contribution, considering the use of low-cost photogrammetric detection methodologies and the use of asset Historical-BIM, has as its aim the theme of knowledge and the adaptation of safety in school buildings, a topic brought to attention by the many situations of seismic risk that have interested the central Apennines in Italy. The specific investigation is referred to the Abruzzo region, hit by the recent earthquakes of 2016 and 2009 that have highlighted the vulnerability of the building structures involved in a large seismic crater covering large areas of the territory. The need to consider in advance the performance standards of building components, especially concerning the strategic ways of the functions contained in them, starts here. In this sense, the school buildings have emerged among the types on which to pay attention, a study theme to be promptly considered, considering the functions performed within them and the possible criticality of such constructions, often dated, enlarged or readjusted without appropriate seismic adaptation plans. From here derives the purpose of the research that is directed towards a systematic recognition of the scholastic heritage, deriving from objective and rapid surveys at low cost, taking into consideration the as-built and the different formal and structural aspects that define the architectural organisms to analyse and manage through three-dimensional models that can be interrogated using HBIM connected to databases containing information of a structural and functional nature. In summary, through the implementation of information in the BIM model, it will be possible to query and obtain in real time all the necessary information to optimize, in terms of efficiency, costs, and future maintenance operations
A comparison of efficient permutation tests for unbalanced ANOVA in two by two designs--and their behavior under heteroscedasticity
We compare different permutation tests and some parametric counterparts that
are applicable to unbalanced designs in two by two designs. First the different
approaches are shortly summarized. Then we investigate the behavior of the
tests in a simulation study. A special focus is on the behavior of the tests
under heteroscedastic variances.Comment: 20 pages, 9 figures, Working Paper of the Department of Management
And Enigineering of the University of Padov
Specification and verification of reconfiguration protocols in grid component systems
In this work we present an approach for the formal specification and verification of the reconfiguration protocols in Grid component systems. We consider Fractal, a modular and extensible component model. As a specification tool we invoke a specific temporal language, separated clausal normal form, which has been shown to be capable of expressing any ECTL+ expression thus, we are able to express the complex fairness properties of a component system. The structure of the normal enables us to directly apply the deductive verification technique, temporal resolution defined in the framework of branching-time temporal logic
Momentum space saturation model for deep inelastic scattering and single inclusive hadron production
We show how the AGBS model, originally developed for deep inelastic
scattering applied to HERA data on the proton structure function, can also
describe the RHIC data on single inclusive hadron yield for and
collisions through a new simultaneous fit. The single inclusive hadron
production is modeled through the color glass condensate, which uses the
quark(and gluon)--condensate amplitudes in momentum space. The AGBS model is
also a momentum space model based on the asymptotic solutions of the BK
equation, although a different definition of the Fourier transform is used.
This aspect is overcome and a description entirely in transverse momentum of
both processes arises for the first time. The small difference between the
simultaneous fit and the one for HERA data alone suggests that the AGBS model
describes very well both kind of processes and thus emerges as a good tool to
investigate the inclusive hadron production data. We use this model for
predictions at LHC energies, which agree very well with available experimental
data.Comment: 10 pages, 7 figure
Is demagnetization an efficient optimization method?
Demagnetization, commonly employed to study ferromagnets, has been proposed
as the basis for an optimization tool, a method to find the ground state of a
disordered system. Here we present a detailed comparison between the ground
state and the demagnetized state in the random field Ising model, combing exact
results in and numerical solutions in . We show that there are
important differences between the two states that persist in the thermodynamic
limit and thus conclude that AC demagnetization is not an efficient
optimization method.Comment: 2 pages, 1 figur
phenomenology at LHC
We study the phenomenology for two extensions of the Electroweak
Standard Model (SM) which have an extra gauge factor. We show the
capabilities of the LHC in distinguishing the signals coming from these two
extensions and both of them from the Standard Model background. In order to
compare the behavior of these models we consider the reaction and compute some observables as the total
cross sections, number of events, forward-backward asymmetry, final particle
distributions like rapidity, transverse momentum, and dimuon invariant mass,
for two LHC regimes: TeV () and
TeV () for = 1000 GeV and 1500
GeV. We show that by using appropriate kinematic cuts some of the observables
considered here are able to extract different properties of the
boson, and hence providing information about to which model it belongs
to.Comment: 21 pages, 17 figures, 4 table
- …