80 research outputs found

    A.P.O. rules are asymptotically non deficient for estimation with squared error loss

    Full text link
    The problem considered is sequential estimation of the mean θ of a one-parameter exponential family of distributions with squared error loss for estimation error and a cost c >0 for each of an i.i.d. sequence of potential observations X 1 , X 2 ,...A Bayesian approach is adopted, and natural conjugate prior distributions are assumed. For this problem, the asymptotically pointwise optimal (A.P.O.) procedure continues sampling until the posterior variance of θ is less than c (r 0 +n), where n is the sample size and r 0 is the fictitous sample size implicit in the conjugate prior distribution. It is known that the A.P.O. procedure is Bayes risk efficient, under mild integrability conditions. In fact, the Bayes risk of both the optimal and A.P.O. procedures are asymptotic to 2 V 0 √c , as c →0, where V 0 is the prior expectation of the standard deviation of X 1 given θ . Here the A.P.O. rule is shown to be asymptotically non-deficient, under stronger regularity conditions: that is, the difference between the Bayes risk of the A.P.O. rule and the Bayes risk of the optimal procedure is of smaller order of magnitude than c , the cost of a single observation, as c →0. The result is illustrated in the exponential and Bernoulli cases, and extended to the case of a normal distribution with both the mean and variance unknown.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/47653/1/440_2004_Article_BF00542639.pd

    Estimation of a probability in inverse binomial sampling under normalized linear-linear and inverse-linear loss

    Get PDF
    Sequential estimation of the success probability pp in inverse binomial sampling is considered in this paper. For any estimator p^\hat p, its quality is measured by the risk associated with normalized loss functions of linear-linear or inverse-linear form. These functions are possibly asymmetric, with arbitrary slope parameters aa and bb for p^p\hat pp respectively. Interest in these functions is motivated by their significance and potential uses, which are briefly discussed. Estimators are given for which the risk has an asymptotic value as pp tends to 00, and which guarantee that, for any pp in (0,1)(0,1), the risk is lower than its asymptotic value. This allows selecting the required number of successes, rr, to meet a prescribed quality irrespective of the unknown pp. In addition, the proposed estimators are shown to be approximately minimax when a/ba/b does not deviate too much from 11, and asymptotically minimax as rr tends to infinity when a=ba=b.Comment: 4 figure

    General Theory of Hypothesis Testing General theory of hypothesis testing

    No full text

    Testing for Ordered Alternatives

    No full text
    • …
    corecore