26,298 research outputs found

    Hidden Dangers to Researcher Safety While Sampling Freshwater Benthic Macroinvertebrates

    Get PDF
    Abstract This paper reviews hidden dangers that threaten the safety of freshwater (FW) researchers of benthic macroinvertebrates (BMIs). Six refereed journals containing 2,075 papers were reviewed for field research resulting in 505 FW BMI articles. However, danger was reported in only 18% of FW BMI papers. I discussed: 1) papers that did not warn of existing danger and consider researcher safety, 2) metric threshold values (e.g., chemical hazards), and non-metric dangers, (e.g., caves and aquatic habitats), 3), the frequency of danger occurrence, 4) baseline and extreme values. Examples of 28 danger factors that posed a threat to BMI researchers in water were compared by frequency per journal papers. FW dangers identified by metric thresholds present a safety limit not to be exceeded, whereas non-metric dangers do not have a threshold as further explained. Also, discussed was a recent thesis on civil engineering hydraulics that identified low-head dams as deceptive and an increasing source of drownings in 39 states. A safe shallow water maximum depth to wade and collect BMIs is proposed based on researcher height and gender, compared to human height means in a large database. Practical safety recommendations were presented to help protect the FW researcher avoid and survive hidden dangers

    A Pedant's Approach to Exponential Smoothing

    Get PDF
    An approach to exponential smoothing that relies on a linear single source of error state space model is outlined. A maximum likelihood method for the estimation of associated smoothing parameters is developed. Commonly used restrictions on the smoothing parameters are rationalised. Issues surrounding model identification and selection are also considered. It is argued that the proposed revised version of exponential smoothing provides a better framework for forecasting than either the Box-Jenkins or the traditional multi-disturbance state space approaches.Time Series Analysis, Prediction, Exponential Smoothing, ARIMA Models, Kalman Filter, State Space Models

    Exponential Smoothing: A Prediction Error Decomposition Principle

    Get PDF
    In the exponential smoothing approach to forecasting, restrictions are often imposed on the smoothing parameters which ensure that certain components are exponentially weighted averages. In this paper, a new general restriction is derived on the basis that the one-step ahead prediction error can be decomposed into permanent and transient components. It is found that this general restriction reduces to the common restrictions used for simple, trend and seasonal exponential smoothing. As such, the prediction error argument provides the rationale for these restrictions.time series analysis, prediction, exponential smoothing, ARIMA models, state space models.

    Spectroscopy of discrete energy levels in ultrasmall metallic grains

    Full text link
    We review recent experimental and theoretical work on ultrasmall metallic grains, i.e. grains sufficiently small that the conduction electron energy spectrum becomes discrete. The discrete excitation spectrum of an individual grain can be measured by the technique of single-electron tunneling spectroscopy: the spectrum is extracted from the current-voltage characteristics of a single-electron transistor containing the grain as central island. We review experiments studying the influence on the discrete spectrum of superconductivity, nonequilibrium excitations, spin-orbit scattering and ferromagnetism. We also review the theoretical descriptions of these phenomena in ultrasmall grains, which require modifications or extensions of the standard bulk theories to include the effects of level discreteness.Comment: 149 pages Latex, 35 figures, to appear in Physics Reports (2001

    Helical Fields and Filamentary Molecular Clouds

    Full text link
    We study the equilibrium of pressure truncated, filamentary molecular clouds that are threaded by rather general helical magnetic fields. We first derive a new virial equation appropriate for magnetized filamentary clouds, which includes the effects of non-thermal motions and the turbulent pressure of the surrounding ISM. When compared with the data, we find that many filamentary clouds have a mass per unit length that is significantly reduced by the effects of external pressure, and that toroidal fields play a significant role in squeezing such clouds. We also develop exact numerical MHD models of filamentary molecular clouds with more general helical field configurations than have previously been considered. We also examine the effects of the equation of state by comparing ``isothermal'' filaments, with constant total (thermal plus turbulent) velocity dispersion, with equilibria constructed using a logatropic equation of state. We perform a Monte Carlo exploration of our parameter space to determine which choices of parameters result in models that agree with the available observational constraints. We find that both equations of state result in equilibria that agree with the observational results. Moreover, we find that models with helical fields have more realistic density profiles than either unmagnetized models or those with purely poloidal fields; we find that most isothermal models have density distributions that fall off as r^{-1.8} to r^{-2}, while logatropes have density profiles that range from r^{-1} to r^{-1.8}. We find that purely poloidal fields produce filaments with steep density gradients that not allowed by the observations.Comment: 21 pages, 8 eps figures, submitted to MNRAS. Significant streamlining of tex

    Reading policies for joins: An asymptotic analysis

    Full text link
    Suppose that mnm_n observations are made from the distribution R\mathbf {R} and nmnn-m_n from the distribution S\mathbf {S}. Associate with each pair, xx from R\mathbf {R} and yy from S\mathbf {S}, a nonnegative score ϕ(x,y)\phi(x,y). An optimal reading policy is one that yields a sequence mnm_n that maximizes E(M(n))\mathbb{E}(M(n)), the expected sum of the (nmn)mn(n-m_n)m_n observed scores, uniformly in nn. The alternating policy, which switches between the two sources, is the optimal nonadaptive policy. In contrast, the greedy policy, which chooses its source to maximize the expected gain on the next step, is shown to be the optimal policy. Asymptotics are provided for the case where the R\mathbf {R} and S\mathbf {S} distributions are discrete and ϕ(x,y)=1or0\phi(x,y)=1 or 0 according as x=yx=y or not (i.e., the observations match). Specifically, an invariance result is proved which guarantees that for a wide class of policies, including the alternating and the greedy, the variable M(n) obeys the same CLT and LIL. A more delicate analysis of the sequence E(M(n))\mathbb{E}(M(n)) and the sample paths of M(n), for both alternating and greedy, reveals the slender sense in which the latter policy is asymptotically superior to the former, as well as a sense of equivalence of the two and robustness of the former.Comment: Published at http://dx.doi.org/10.1214/105051606000000646 in the Annals of Applied Probability (http://www.imstat.org/aap/) by the Institute of Mathematical Statistics (http://www.imstat.org

    A comparison of integrated testlet and constructed-response question formats

    Full text link
    Constructed-response (CR) questions are a mainstay of introductory physics textbooks and exams. However, because of time, cost, and scoring reliability constraints associated with this format, CR questions are being increasingly replaced by multiple-choice (MC) questions in formal exams. The integrated testlet (IT) is a recently-developed question structure designed to provide a proxy of the pedagogical advantages of CR questions while procedurally functioning as set of MC questions. ITs utilize an answer-until-correct response format that provides immediate confirmatory or corrective feedback, and they thus allow not only for the granting of partial credit in cases of initially incorrect reasoning, but furthermore the ability to build cumulative question structures. Here, we report on a study that directly compares the functionality of ITs and CR questions in introductory physics exams. To do this, CR questions were converted to concept-equivalent ITs, and both sets of questions were deployed in midterm and final exams. We find that both question types provide adequate discrimination between stronger and weaker students, with CR questions discriminating slightly better than the ITs. Meanwhile, an analysis of inter-rater scoring of the CR questions raises serious concerns about the reliability of the granting of partial credit when this traditional assessment technique is used in a realistic (but non optimized) setting. Furthermore, we show evidence that partial credit is granted in a valid manner in the ITs. Thus, together with consideration of the vastly reduced costs of administering IT-based examinations compared to CR-based examinations, our findings indicate that ITs are viable replacements for CR questions in formal examinations where it is desirable to both assess concept integration and to reward partial knowledge, while efficiently scoring examinations.Comment: 14 pages, 3 figures, with appendix. Accepted for publication in PRST-PER (August 2014
    corecore