2,024 research outputs found
Recombination Algorithms and Jet Substructure: Pruning as a Tool for Heavy Particle Searches
We discuss jet substructure in recombination algorithms for QCD jets and
single jets from heavy particle decays. We demonstrate that the jet algorithm
can introduce significant systematic effects into the substructure. By
characterizing these systematic effects and the substructure from QCD,
splash-in, and heavy particle decays, we identify a technique, pruning, to
better identify heavy particle decays into single jets and distinguish them
from QCD jets. Pruning removes protojets typical of soft, wide angle radiation,
improves the mass resolution of jets reconstructing a heavy particle decay, and
decreases the QCD background. We show that pruning provides significant
improvements over unpruned jets in identifying top quarks and W bosons and
separating them from a QCD background, and may be useful in a search for heavy
particles.Comment: 33 pages, 42 figure
On Statistical Aspects of Qjets
The process by which jet algorithms construct jets and subjets is inherently
ambiguous and equally well motivated algorithms often return very different
answers. The Qjets procedure was introduced by the authors to account for this
ambiguity by considering many reconstructions of a jet at once, allowing one to
assign a weight to each interpretation of the jet. Employing these weighted
interpretations leads to an improvement in the statistical stability of many
measurements. Here we explore in detail the statistical properties of these
sets of weighted measurements and demonstrate how they can be used to improve
the reach of jet-based studies.Comment: 29 pages, 6 figures. References added, minor modification of the
text. This version to appear in JHE
Qjets: A Non-Deterministic Approach to Tree-Based Jet Substructure
Jet substructure is typically studied using clustering algorithms, such as
kT, which arrange the jets' constituents into trees. Instead of considering a
single tree per jet, we propose that multiple trees should be considered,
weighted by an appropriate metric. Then each jet in each event produces a
distribution for an observable, rather than a single value. Advantages of this
approach include: 1) observables have significantly increased statistical
stability; and, 2) new observables, such as the variance of the distribution,
provide new handles for signal and background discrimination. For example, we
find that employing a set of trees substantially reduces the observed
fluctuations in the pruned mass distribution, enhancing the likelihood of new
particle discovery for a given integrated luminosity. Furthermore, the
resulting pruned mass distributions for (background) QCD jets are found to be
substantially wider than that for (signal) jets with intrinsic mass scales,
e.g. jets containing a W decay. A cut on this width yields a substantial
enhancement in significance relative to a cut on the standard pruned jet mass
alone. In particular the luminosity needed for a given significance requirement
decreases by a factor of two relative to standard pruning.Comment: Minor changes to match journal versio
QCD corrections to stoponium production at hadron colliders
If the lighter top squark has no kinematically allowed two-body decays that
conserve flavor, then it will live long enough to form hadronic bound states.
The observation of the diphoton decays of stoponium could then provide a
uniquely precise measurement of the top squark mass. In this paper, we
calculate the cross section for the production of stoponium in a hadron
collider at next-to-leading order (NLO) in QCD. We present numerical results
for the cross section for production of stoponium at the LHC and study the
dependence on beam energy, stoponium mass, and the renormalization and
factorization scale. The cross-section is substantially increased by the NLO
corrections, counteracting a corresponding decrease found earlier in the NLO
diphoton branching ratio.Comment: 24 page
Successive Combination Jet Algorithm For Hadron Collisions
Jet finding algorithms, as they are used in and hadron collisions,
are reviewed and compared. It is suggested that a successive combination style
algorithm, similar to that used in physics, might be useful also in
hadron collisions, where cone style algorithms have been used previously.Comment: 18 pages plus four uuencoded postscript figures, REVTEX 3.0,
CERN-TH.6860/9
Visual Stability of Objects and Environments Viewed through Head-Mounted Displays
Virtual Environments (aka Virtual Reality) is again catching the public imagination and a number of startups (e.g. Oculus) and even not-so-startup companies (e.g. Microsoft) are trying to develop display systems to capitalize on this renewed interest. All acknowledge that this time they will get it right by providing the required dynamic fidelity, visual quality, and interesting content for the concept of VR to take off and change the world in ways it failed to do so in past incarnations. Some of the surprisingly long historical background of the technology that the form of direct simulation that underlies virtual environment and augmented reality displays will be briefly reviewed. An example of a mid 1990's augmented reality display system with good dynamic performance from our lab will be used to illustrate some of the underlying phenomena and technology concerning visual stability of virtual environments and objects during movement. In conclusion some idealized performance characteristics for a reference system will be proposed. Interestingly, many systems more or less on the market now may actually meet many of these proposed technical requirements. This observation leads to the conclusion that the current success of the IT firms trying to commercialize the technology will depend on the hidden costs of using the systems as well as the development of interesting and compelling content
- …