10,247 research outputs found
Incidence, severity and prognosis associated with hypernatremia in dogs and cats.
BackgroundHypernatremia has been associated with substantial morbidity and death in human patients. The incidence and importance of hypernatremia in dogs and cats has not been determined.Hypothesis/objectivesTo describe the incidence of and prognosis associated with hypernatremia in dogs and cats at a university teaching hospital.AnimalsA total of 16,691 dogs and 4,211 cats with measured blood or serum sodium concentration.MethodsRetrospective study. Medical records of animals with a blood or serum sodium concentration measured during a 60-month period were reviewed to determine the severity of hypernatremia and its associated case fatality rate. Cases with moderate (11-15 mmol/L above the reference range) or severe hypernatremia (≥16 mmol/L above the reference range) were further reviewed.ResultsA total of 957 dogs (5.7%) and 338 cats (8.0%) were diagnosed with hypernatremia. Case fatality rates of dogs and cats with hypernatremia was 20.6 and 28.1%, respectively compared to 4.4 and 4.5% with a normal blood or serum sodium concentration (P < .0001). The magnitude of hypernatremia was linearly associated with a higher case fatality rate (P < .0001). Hypernatremia was associated with a higher case fatality rate than hyponatremia. Among the animals with moderate or severe hypernatremia, 50% of dogs and 38.5% of cats presented with community-acquired hypernatremia, and 50% of dogs and 61.5% of cats developed hospital-acquired hypernatremia.Conclusions and clinical importanceHypernatremia was found infrequently in this population but was associated with increased case fatality rates in dogs and cats. Presence and severity of hypernatremia might be useful as a prognostic indicator
Prediction of lactate threshold using the modified Conconi test in distance runners
This study aimed to examine the validity of the modified Conconi test (CT) to predict lactate threshold (LT) during running. Twelve distance runners randomly performed the modified CT and the incremental test in which LT was determined directly by measuring blood lactate (BLa). Mean values of heart rate (HR) and running speed (RS) at heart rate deflection point (HRDP) obtained through the modified CT were compared with those at LT. Subsequently, the runners who showed a HRDP in the modified CT performed a 30-min prolonged exercise test (PET) at a RS corresponding to HRDP. During this test, the kinetics of BLa and HR were analyzed to determine whether a steady state in these variables could be attained. We succeeded in identifying HRDP in nine of our runners, whereas the remaining three runners showed entirely linear HR response. In those nine runners, no significant difference was found between HR and RS at HRDP and those at LT. Significant correlation was found between HR at HRDP and HR at LT (r = 0.84, p = 0.005), but RS at HRDP was not significantly correlated with RS at LT (r = 0.63, p = 0.07). All nine runners were able to complete the PET with steady state conditions being achieved for both BLa and HR. In conclusion, these findings indicate that the modified CT has a potential to be used as an alternative method for assessment of LT in distance runners presenting a HRDP
Thermoelectric Amplification of Phonons in Graphene
Amplification of acoustic phonons due to an external temperature gredient
() in Graphene was studied theoretically. The threshold temperature
gradient at which absorption switches over to amplification
in Graphene was evaluated at various frequencies and temperatures
. For and frequency , . The calculation was done in the regime at . The
dependence of the normalized () on the frequency
and the temperature gradient are evaluated numerically and
presented graphically. The calculated for Graphene is lower
than that obtained for homogeneous semiconductors () , Superlattices , Cylindrical Quantum Wire . This makes Graphene a much better material for thermoelectric
phonon amplifier.Comment: 12 Pages, 6 figure
The Emergence of Norms via Contextual Agreements in Open Societies
This paper explores the emergence of norms in agents' societies when agents
play multiple -even incompatible- roles in their social contexts
simultaneously, and have limited interaction ranges. Specifically, this article
proposes two reinforcement learning methods for agents to compute agreements on
strategies for using common resources to perform joint tasks. The computation
of norms by considering agents' playing multiple roles in their social contexts
has not been studied before. To make the problem even more realistic for open
societies, we do not assume that agents share knowledge on their common
resources. So, they have to compute semantic agreements towards performing
their joint actions. %The paper reports on an empirical study of whether and
how efficiently societies of agents converge to norms, exploring the proposed
social learning processes w.r.t. different society sizes, and the ways agents
are connected. The results reported are very encouraging, regarding the speed
of the learning process as well as the convergence rate, even in quite complex
settings
Vacuum polarization around stars: nonlocal approximation
We compute the vacuum polarization associated with quantum massless fields
around stars with spherical symmetry. The nonlocal contribution to the vacuum
polarization is dominant in the weak field limit, and induces quantum
corrections to the exterior metric that depend on the inner structure of the
star. It also violates the null energy conditions. We argue that similar
results also hold in the low energy limit of quantum gravity. Previous
calculations of the vacuum polarization in spherically symmetric spacetimes,
based on local approximations, are not adequate for newtonian stars.Comment: 8 pages, no figure
Resource augmentation in load balancing
We consider load-balancing in the following setting. The on-line algorithm is allowed to use machines, whereas the optimal off-line algorithm is limited to machines, for some fixed . We show that while the greedy algorithm has a competitive ratio which decays linearly in the inverse of , the best on-line algorithm has a ratio which decays exponentially in . Specifically, we give an algorithm with competitive ratio of 1+2^{- frac{n{m (1- o (1)), and a lower bound of 1+ e^{ - frac{n{m (1+ o(1)) on the competitive ratio of any randomized algorithm. We also consider the preemptive case. We show an on-line algorithm with a competitive ratio of 1+ e^{ - frac{n{m (1+ o(1)). We show that the algorithm is optimal by proving a matching lower bound. We also consider the non-preemptive model with temporary tasks. We prove that for , the greedy algorithm is optimal. (It is not optimal for permanent tasks.
Explicit Zeta Functions for Bosonic and Fermionic Fields on a Noncommutative Toroidal Spacetime
Explicit formulas for the zeta functions corresponding to
bosonic () and to fermionic () quantum fields living on a
noncommutative, partially toroidal spacetime are derived. Formulas for the most
general case of the zeta function associated to a quadratic+linear+constant
form (in {\bf Z}) are obtained. They provide the analytical continuation of the
zeta functions in question to the whole complex plane, in terms of series
of Bessel functions (of fast, exponential convergence), thus being extended
Chowla-Selberg formulas. As well known, this is the most convenient expression
that can be found for the analytical continuation of a zeta function, in
particular, the residua of the poles and their finite parts are explicitly
given there. An important novelty is the fact that simple poles show up at
, as well as in other places (simple or double, depending on the number of
compactified, noncompactified, and noncommutative dimensions of the spacetime),
where they had never appeared before. This poses a challenge to the
zeta-function regularization procedure.Comment: 15 pages, no figures, LaTeX fil
Rank-Ordering Statistics of Extreme Events: Application to the Distribution of Large Earthquakes
Rank-ordering statistics provides a perspective on the rare, largest elements
of a population, whereas the statistics of cumulative distributions are
dominated by the more numerous small events. The exponent of a power law
distribution can be determined with good accuracy by rank-ordering statistics
from the observation of only a few tens of the largest events. Using analytical
results and synthetic tests, we quantify the systematic and the random errors.
We also study the case of a distribution defined by two branches, each having
a power law distribution, one defined for the largest events and the other for
smaller events, with application to the World-Wide (Harvard) and Southern
California earthquake catalogs. In the case of the Harvard moment catalog, we
make more precise earlier claims of the existence of a transition of the
earthquake magnitude distribution between small and large earthquakes; the
-values are for large shallow earthquakes and for smaller shallow earthquakes. However, the cross-over
magnitude between the two distributions is ill-defined. The data available at
present do not provide a strong constraint on the cross-over which has a
probability of being between magnitudes and for shallow
earthquakes; this interval may be too conservatively estimated. Thus, any
influence of a universal geometry of rupture on the distribution of earthquakes
world-wide is ill-defined at best. We caution that there is no direct evidence
to confirm the hypothesis that the large-moment branch is indeed a power law.
In fact, a gamma distribution fits the entire suite of earthquake moments from
the smallest to the largest satisfactorily. There is no evidence that the
earthquakes of the Southern California catalog have a distribution with tw
Monte Carlo Simulations for Ghost Imaging Based on Scattered Photons
X-ray based imaging modalities are widely used in research, industry, and in
the medical field. Consequently, there is a strong motivation to improve their
performances with respect to resolution, dose, and contrast. Ghost imaging (GI)
is an imaging technique in which the images are reconstructed from measurements
with a single-pixel detector using correlation between the detected intensities
and the intensity structures of the input beam. The method that has been
recently extended to X-rays provides intriguing possibilities to overcome
several fundamental challenges of X-ray imaging. However, understanding the
potential of the method and designing X-ray GI systems pose challenges since in
addition to geometric optic effects, radiation-matter interactions must be
considered. Such considerations are fundamentally more complex than those at
longer wavelengths as relativistic effects such as Compton scattering become
significant. In this work we present a new method for designing and
implementing GI systems using the particle transport code FLUKA, that rely on
Monte Carlo (MC) sampling. This new approach enables comprehensive
consideration of the radiation-matter interactions, facilitating successful
planning of complex GI systems. As an example of an advanced imaging system, we
simulate a high-resolution scattered photons GI technique
- …