17,546 research outputs found

    Optimizing passive acoustic sampling of bats in forests

    Get PDF
    Passive acoustic methods are increasingly used in biodiversity research and monitoring programs because they are cost-effective and permit the collection of large datasets. However, the accuracy of the results depends on the bioacoustic characteristics of the focal taxa and their habitat use. In particular, this applies to bats which exhibit distinct activity patterns in three-dimensionally structured habitats such as forests. We assessed the performance of 21 acoustic sampling schemes with three temporal sampling patterns and seven sampling designs. Acoustic sampling was performed in 32 forest plots, each containing three microhabitats: forest ground, canopy, and forest gap. We compared bat activity, species richness, and sampling effort using species accumulation curves fitted with the clench equation. In addition, we estimated the sampling costs to undertake the best sampling schemes. We recorded a total of 145,433 echolocation call sequences of 16 bat species. Our results indicated that to generate the best outcome, it was necessary to sample all three microhabitats of a given forest location simultaneously throughout the entire night. Sampling only the forest gaps and the forest ground simultaneously was the second best choice and proved to be a viable alternative when the number of available detectors is limited. When assessing bat species richness at the 1-km(2) scale, the implementation of these sampling schemes at three to four forest locations yielded highest labor cost-benefit ratios but increasing equipment costs. Our study illustrates that multiple passive acoustic sampling schemes require testing based on the target taxa and habitat complexity and should be performed with reference to cost-benefit ratios. Choosing a standardized and replicated sampling scheme is particularly important to optimize the level of precision in inventories, especially when rare or elusive species are expected

    Mismeasured Mortality: Correcting Estimates of Wolf Poaching in the United States

    Get PDF
    Measuring rates and causes of mortalities is important in animal ecology and management. Observing the fates of known individuals is a common method of estimating life history variables, including mortality patterns. It has long been assumed that data lost when known animals disappear were unbiased. We test and reject this assumption under conditions common to most, if not all, studies using marked animals. We illustrate the bias for 4 endangered wolf populations in the United States by reanalyzing data and assumptions about the known and unknown fates of marked wolves to calculate the degree to which risks of different causes of death were mismeasured. We find that, when using traditional methods, the relative risk of mortality from legal killing measured as a proportion of all known fates was overestimated by 0.05–0.16 and the relative risk of poaching was underestimated by 0.17–0.44. We show that published government estimates are affected by these biases and, importantly, are underestimating the risk of poaching. The underestimates have obscured the magnitude of poaching as the major threat to endangered wolf populations. We offer methods to correct estimates of mortality risk for marked animals of any taxon and describe the conditions under which traditional methods produce more or less bias. We also show how correcting past and future estimates of mortality parameters can address uncertainty about wildlife populations and increase the predictability and sustainability of wildlife management interventions

    The First Science Results from SPHERE: Disproving the Predicted Brown Dwarf around V471 Tau

    Get PDF
    Variations of eclipse arrival times have recently been detected in several post common envelope binaries consisting of a white dwarf and a main sequence companion star. The generally favoured explanation for these timing variations is the gravitational pull of one or more circumbinary substellar objects periodically moving the center of mass of the host binary. Using the new extreme-AO instrument SPHERE, we image the prototype eclipsing post-common envelope binary V471 Tau in search of the brown dwarf that is believed to be responsible for variations in its eclipse arrival times. We report that an unprecedented contrast of 12.1 magnitudes in the H band at a separation of 260 mas was achieved, but resulted in a non-detection. This implies that there is no brown dwarf present in the system unless it is three magnitudes fainter than predicted by evolutionary track models, and provides damaging evidence against the circumbinary interpretation of eclipse timing variations. In the case of V471 Tau, a more consistent explanation is offered with the Applegate mechanism, in which these variations are prescribed to changes in the quadrupole moment within the main-sequence sta

    Cultural and economic complementarities of spatial agglomeration in the British television broadcasting industry: Some explorations.

    Get PDF
    This paper considers the processes supporting agglomeration in the British television broadcasting industry. It compares and contrasts the insights offered by the cultural turn in geography and more conventionally economic approaches. It finds that culture and institutions are fundamental to the constitution of production and exchange relationships and also that they solve fundamental economic problems of coordinating resources under conditions of uncertainty and limited information. Processes at a range of spatial scales are important, from highly local to global, and conventional economics casts some light on which firms are most active and successful

    Early multi-wavelength emission from Gamma-ray Bursts: from Gamma-ray to X-ray

    Get PDF
    The study of the early high-energy emission from both long and short Gamma-ray bursts has been revolutionized by the Swift mission. The rapid response of Swift shows that the non-thermal X-ray emission transitions smoothly from the prompt phase into a decaying phase whatever the details of the light curve. The decay is often categorized by a steep-to-shallow transition suggesting that the prompt emission and the afterglow are two distinct emission components. In those GRBs with an initially steeply-decaying X-ray light curve we are probably seeing off-axis emission due to termination of intense central engine activity. This phase is usually followed, within the first hour, by a shallow decay, giving the appearance of a late emission hump. The late emission hump can last for up to a day, and hence, although faint, is energetically very significant. The energy emitted during the late emission hump is very likely due to the forward shock being constantly refreshed by either late central engine activity or less relativistic material emitted during the prompt phase. In other GRBs the early X-ray emission decays gradually following the prompt emission with no evidence for early temporal breaks, and in these bursts the emission may be dominated by classical afterglow emission from the external shock as the relativistic jet is slowed by interaction with the surrounding circum-burst medium. At least half of the GRBs observed by Swift also show erratic X-ray flaring behaviour, usually within the first few hours. The properties of the X-ray flares suggest that they are due to central engine activity. Overall, the observed wide variety of early high-energy phenomena pose a major challenge to GRB models.Comment: Accepted for publication in the New Journal of Physics focus issue on Gamma Ray Burst

    Redundant interferometric calibration as a complex optimization problem

    Get PDF
    Observations of the redshifted 21-cm line from the epoch of reionization have recently motivated the construction of low frequency radio arrays with highly redundant configurations. These configurations provide an alternative calibration strategy - "redundant calibration" - and boosts sensitivity on specific spatial scales. In this paper, we formulate calibration of redundant interferometric arrays as a complex optimization problem. We solve this optimization problem via the Levenberg-Marquardt algorithm. This calibration approach is more robust to initial conditions than current algorithms and, by leveraging an approximate matrix inversion, allows for further optimization and an efficient implementation ("redundant StEfCal"). We also investigated using the preconditioned conjugate gradient method as an alternative to the approximate matrix inverse, but found that its computational performance is not competitive with respect to "redundant StEfCal". The efficient implementation of this new algorithm is made publicly available.Comment: 11 pages, 7 figures, MNRAS accepte

    Cavitation inception of a van der Waals fluid at a sack-wall obstacle

    Full text link
    Cavitation in a liquid moving past a constraint is numerically investigated by means of a free-energy lattice Boltzmann simulation based on the van der Waals equation of state. The fluid is streamed past an obstacle and, depending on the pressure drop between inlet and outlet, vapor formation underneath the corner of the sack-wall is observed. The circumstances of cavitation formation are investigated and it is found that the local bulk pressure and mean stress are insufficient to explain the phenomenon. Results obtained in this study strongly suggest that the viscous stress, interfacial contributions to the local pressure, and the Laplace pressure are relevant to the opening of a vapor cavity. This can be described by a generalization of Joseph's criterion that includes these contributions. A macroscopic investigation measuring mass flow rate behavior and discharge coefficient was also performed. As theoretically predicted, mass flow rate increases linearly with the square root of the pressure drop. However, when cavitation occurs, the mass flow growth rate is reduced and eventually it collapses into a choked flow state. In the cavitating regime, as theoretically predicted and experimentally verified, the discharge coefficient grows with the Nurick cavitation number

    Influence of Supercurrents on Low-Temperature Thermopower in Mesoscopic N/S Structures

    Full text link
    The thermopower of mesoscopic normal metal/superconductor structures has been measured at low temperatures. Effect of supercurrent present in normal part of the structure was studied in two cases: when it was created by applied external magnetic field and when it was applied directly using extra superconducting electrodes. Temperature and magnetic field dependencies of thermopower are compared to the numerical simulations based on the quasiclassical theory of the superconducting proximity effect.Comment: 21 pages, 12 figures. To be published in the proceedings of the ULTI conference organized in Lammi, Finland (2006

    Confirmation of the \eps -- \eiso (Amati) relation from the X-ray flash XRF 050416A observed by Swift/BAT

    Full text link
    We report Swift Burst Alert Telescope (BAT) observations of the X-ray Flash (XRF) XRF 050416A. The fluence ratio between the 15-25 keV and 25-50 keV energy bands of this event is 1.5, thus making it the softest gamma-ray burst (GRB) observed by BAT so far. The spectrum is well fitted by the Band function with E^{\rm obs}_{\rm peak} of 15.0_{-2.7}^{+2.3} keV. Assuming the redshift of the host galaxy (z = 0.6535), the isotropic-equivalent radiated energy E_{\rm iso} and the peak energy at the GRB rest frame (E^{\rm src}_{\rm peak}) of XRF 050416A are not only consistent with the correlation found by Amati et al. and extended to XRFs by Sakamoto et al., but also fill-in the gap of this relation around the 30 - 80 keV range of E^{\rm src}_{\rm peak}. This result tightens the validity of the E^{\rm src}_{\rm peak} - E_{\rm iso} relation from XRFs to GRBs. We also find that the jet break time estimated using the empirical relation between E^{\rm src}_{\rm peak} and the collimation corrected energy E_{\gamma} is inconsistent with the afterglow observation by Swift X-ray Telescope. This could be due to the extra external shock emission overlaid around the jet break time or to the non existence of a jet break feature for XRF, which might be a further challenging for GRB jet emission, models and XRF/GRB unification scenarios.Comment: 16 pages, 4 figures; accepted for publication in ApJ

    The Non-linear Dynamics of Meaning-Processing in Social Systems

    Full text link
    Social order cannot be considered as a stable phenomenon because it contains an order of reproduced expectations. When the expectations operate upon one another, they generate a non-linear dynamics that processes meaning. Specific meaning can be stabilized, for example, in social institutions, but all meaning arises from a horizon of possible meanings. Using Luhmann's (1984) social systems theory and Rosen's (1985) theory of anticipatory systems, I submit equations for modeling the processing of meaning in inter-human communication. First, a self-referential system can use a model of itself for the anticipation. Under the condition of functional differentiation, the social system can be expected to entertain a set of models; each model can also contain a model of the other models. Two anticipatory mechanisms are then possible: one transversal between the models, and a longitudinal one providing the modeled systems with meaning from the perspective of hindsight. A system containing two anticipatory mechanisms can become hyper-incursive. Without making decisions, however, a hyper-incursive system would be overloaded with uncertainty. Under this pressure, informed decisions tend to replace the "natural preferences" of agents and an order of cultural expectations can increasingly be shaped
    corecore