5,302 research outputs found
The leptonic decay using the principle of maximum conformality
In the paper, we study the leptonic decay width
by using the principle of maximum
conformality (PMC) scale-setting approach. The PMC adopts the renormalization
group equation to set the correct momentum flow of the process, whose value is
independent to the choice of the renormalization scale and its prediction thus
avoids the conventional renormalization scale ambiguities. Using the known
next-to-next-to-next-to-leading order perturbative series together with the PMC
single scale-setting approach, we do obtain a renormalization scale independent
decay width, keV,
where the error is squared average of those from
, GeV and the choices of
factorization scales within of their central values. To compare with
the result under conventional scale-setting approach, this decay width agrees
with the experimental value within errors, indicating the importance of a
proper scale-setting approach.Comment: 6 pages, 4 figure
Is the late near-infrared bump in short-hard GRB 130603B due to the Li-Paczynski kilonova?
Short-hard gamma-ray bursts (GRBs) are widely believed to be produced by the
merger of two binary compact objects, specifically by two neutron stars or by a
neutron star orbiting a black hole. According to the Li-Paczynski kilonova
model, the merger would launch sub-relativistic ejecta and a
near-infrared/optical transient would then occur, lasting up to days, which is
powered by the radioactive decay of heavy elements synthesized in the ejecta.
The detection of a late bump using the {\em Hubble Space Telescope} ({\em HST})
in the near-infrared afterglow light curve of the short-hard GRB 130603B is
indeed consistent with such a model. However, as shown in this Letter, the
limited {\em HST} near-infrared lightcurve behavior can also be interpreted as
the synchrotron radiation of the external shock driven by a wide mildly
relativistic outflow. In such a scenario, the radio emission is expected to
peak with a flux of Jy, which is detectable for current radio
arrays. Hence, the radio afterglow data can provide complementary evidence on
the nature of the bump in GRB 130603B. It is worth noting that good
spectroscopy during the bump phase in short-hard bursts can test validity of
either model above, analogous to spectroscopy of broad-lined Type Ic supernova
in long-soft GRBs.Comment: 4 pages, 2 figures, published in ApJ Lette
Pneumonia and ocular disease as the primary presentations of Takayasu arteritis: a case report
Analysis of the Influence of Fish Behavior on the Hydrodynamics of Net Cage
In net cage hydrodynamic analysis, drag force of net is dependent on the physical dimensions of the net cage, the Solidity ratio, the Reynolds number and the projected area of the net, which is illustrated in numerous previous researches. However, rare studies attempt to investigate the effect of fish behavior. Thus a net-fluid interaction model and a simplified fish model were proposed for analyzing the effects of fish behavior on the net cage. A series of physical model tests were conducted to validate the numerical model, which indicates models can simulate the stocked net cage in the current accurately. The simulation results indicate that circular movement of fish leads to a low pressure zone at the center of net cage, which causes a strong vertical flow along the center line of the net cage. The drag force on the net cage is significantly decreased with the increasing fish stocking density
Anisotropic Surface Remeshing without Obtuse Angles
We present a novel anisotropic surface remeshing method that can efficiently eliminate obtuse angles. Unlike previous work that can only suppress obtuse angles with expensive resampling and Lloyd-type iterations, our method relies on a simple yet efficient connectivity and geometry refinement, which can not only remove all the obtuse angles, but also preserves the original mesh connectivity as much as possible. Our method can be directly used as a post-processing step for anisotropic meshes generated from existing algorithms to improve mesh quality. We evaluate our method by testing on a variety of meshes with different geometry and topology, and comparing with representative prior work. The results demonstrate the effectiveness and efficiency of our approach
- …