793 research outputs found

    The Chernoff Modification of the Fisher Exact Test for the Difference of Two Binomial Probabilities

    Get PDF
    To test the difference of two binomial probabilities, we will apply the Fisher exact test modified by Chernoff (2002), which will be called CMF test. The key idea is to use the P-value of Fisher exact test as a test statistic and this makes it possible to apply the Fisher exact test for more general problems. The numerical comparison with the likelihood ratio test and the Bayes test are also conducted and we find that the performance of CMF test is close to that of the likelihood ratio test and it is satisfactory in various situations. We also give a Kullback-Leibler information number of the margins to for the testing problems which justify the stability of P-values of CMF test.Fisher Exact Test, Binomial Probabilities, Chernoff Modification, Test of Difference of Probabilities

    Modelling network travel time reliability under stochastic demand

    Get PDF
    A technique is proposed for estimating the probability distribution of total network travel time, in the light of normal day-to-day variations in the travel demand matrix over a road traffic network. A solution method is proposed, based on a single run of a standard traffic assignment model, which operates in two stages. In stage one, moments of the total travel time distribution are computed by an analytic method, based on the multivariate moments of the link flow vector. In stage two, a flexible family of density functions is fitted to these moments. It is discussed how the resulting distribution may in practice be used to characterise unreliability. Illustrative numerical tests are reported on a simple network, where the method is seen to provide a means for identifying sensitive or vulnerable links, and for examining the impact on network reliability of changes to link capacities. Computational considerations for large networks, and directions for further research, are discussed

    Global permutation tests for multivariate ordinal data: alternatives, test statistics, and the null dilemma

    Get PDF
    We discuss two-sample global permutation tests for sets of multivariate ordinal data in possibly high-dimensional setups, motivated by the analysis of data collected by means of the World Health Organisation's International Classification of Functioning, Disability and Health. The tests do not require any modelling of the multivariate dependence structure. Specifically, we consider testing for marginal inhomogeneity and direction-independent marginal order. Max-T test statistics are known to lead to good power against alternatives with few strong individual effects. We propose test statistics that can be seen as their counterparts for alternatives with many weak individual effects. Permutation tests are valid only if the two multivariate distributions are identical under the null hypothesis. By means of simulations, we examine the practical impact of violations of this exchangeability condition. Our simulations suggest that theoretically invalid permutation tests can still be 'practically valid'. In particular, they suggest that the degree of the permutation procedure's failure may be considered as a function of the difference in group-specific covariance matrices, the proportion between group sizes, the number of variables in the set, the test statistic used, and the number of levels per variable

    Lectures on Statistics in Theory: Prelude to Statistics in Practice

    Full text link
    This is a writeup of lectures on "statistics" that have evolved from the 2009 Hadron Collider Physics Summer School at CERN to the forthcoming 2018 school at Fermilab. The emphasis is on foundations, using simple examples to illustrate the points that are still debated in the professional statistics literature. The three main approaches to interval estimation (Neyman confidence, Bayesian, likelihood ratio) are discussed and compared in detail, with and without nuisance parameters. Hypothesis testing is discussed mainly from the frequentist point of view, with pointers to the Bayesian literature. Various foundational issues are emphasized, including the conditionality principle and the likelihood principle.Comment: 93 pages, new appendix on goodness of fi

    A comparison of political violence by left-wing, right-wing, and Islamist extremists in the United States and the world

    Get PDF
    Although political violence has been perpetrated on behalf of a wide range of political ideologies it is unclear whether there are systematic differences between ideologies in the use of violence to pursue a political cause. Prior research on this topic is scarce and mostly restricted to self- reported measures or less extreme forms of political aggression. Moreover, it has generally focused on respondents in western countries and has been limited to either comparisons of the supporters of left-wing and right-wing causes or examinations of only Islamist extremism. In this research we address these gaps by comparing the use of political violence by left-wing, right- wing, and Islamist extremists in the United States and worldwide using two unique datasets that cover real-world examples of politically motivated, violent behaviors. Across both datasets, we find that radical acts perpetrated by individuals associated with left-wing causes are less likely to be violent. In the United States we found no difference between the level of violence perpetrated by right-wing and Islamist extremists. However, differences in violence emerged on the global level with Islamist extremists being more likely than right-wing extremists to engage in more violent acts

    Accurate Parametric Inference for Small Samples

    Get PDF
    We outline how modern likelihood theory, which provides essentially exact inferences in a variety of parametric statistical problems, may routinely be applied in practice. Although the likelihood procedures are based on analytical asymptotic approximations, the focus of this paper is not on theory but on implementation and applications. Numerical illustrations are given for logistic regression, nonlinear models, and linear non-normal models, and we describe a sampling approach for the third of these classes. In the case of logistic regression, we argue that approximations are often more appropriate than `exact' procedures, even when these exist.Comment: Published in at http://dx.doi.org/10.1214/08-STS273 the Statistical Science (http://www.imstat.org/sts/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Should We Test the Model Assumptions Before Running a Model-based Test?

    Get PDF
    Statistical methods are based on model assumptions, and it is statistical folklore that a method's model assumptions should be checked before applying it. This can be formally done by running one or more misspecification tests testing model assumptions before running a method that requires these assumptions; here we focus on model-based tests. A combined test procedure can be defined by specifying a protocol in which first model assumptions are tested and then, conditionally on the outcome, a test is run that requires or does not require the tested assumptions. Although such an approach is often taken in practice, much of the literature that investigated this is surprisingly critical of it. Our aim is to explore conditions under which model checking is advisable or not advisable. For this, we review results regarding such ``combined procedures'' in the literature, we review and discuss controversial views on the role of model checking in statistics, and we present a general setup in which we can show that preliminary model checking is advantageous, which implies conditions for making model checking worthwhile

    Divergence vs. Decision P-values: A Distinction Worth Making in Theory and Keeping in Practice

    Full text link
    There are two distinct definitions of 'P-value' for evaluating a proposed hypothesis or model for the process generating an observed dataset. The original definition starts with a measure of the divergence of the dataset from what was expected under the model, such as a sum of squares or a deviance statistic. A P-value is then the ordinal location of the measure in a reference distribution computed from the model and the data, and is treated as a unit-scaled index of compatibility between the data and the model. In the other definition, a P-value is a random variable on the unit interval whose realizations can be compared to a cutoff alpha to generate a decision rule with known error rates under the model and specific alternatives. It is commonly assumed that realizations of such decision P-values always correspond to divergence P-values. But this need not be so: Decision P-values can violate intuitive single-sample coherence criteria where divergence P-values do not. It is thus argued that divergence and decision P-values should be carefully distinguished in teaching, and that divergence P-values are the relevant choice when the analysis goal is to summarize evidence rather than implement a decision rule.Comment: 49 pages. Scandinavian Journal of Statistics 2023, issue 1, with discussion and rejoinder in issue

    Density forecasting in financial risk modelling

    Get PDF
    As a result of an increasingly stringent regulation aimed at monitoring financial risk exposures, nowadays the risk measurement systems play a crucial role in all banks. In this thesis we tackle a variety of problems, related to density forecasting, which are fundamental to market risk managers. The computation of risk measures (e.g. Value-at-Risk) for any portfolio of financial assets requires the generation of density forecasts for the driving risk factors. Appropriate testing procedures must then be identified for an accurate appraisal of these forecasts. We start our research by assessing whether option-implied densities, which constitute the most obvious forecasts of the distribution of the underlying asset at expiry, do actually represent unbiased forecasts. We first extract densities from options on currency and equity index futures, by means of both traditional and original specifications. We then appraise them, via rigorous density forecast evaluation tools, and we find evidence of the presence of biases. In the second part of the thesis, we focus on modelling the dynamics of the volatility curve, in order to measure the vega risk exposure for various delta-hedged option portfolios. We propose to use a linear Kalman filter approach, which gives more precise forecasts of the vega risk exposure than alternative, well-established models. In the third part, we derive a continuous time model for the dynamics of equity index returns from a data set of 5-minute returns. A model inferred from high-frequency typical of risk measures calculations. The last part of our work deals with evaluating density forecasts of the joint distribution of the risk factors. We find that, given certain specifications for the multivariate density forecast, a goodness-of-fit procedure based on the Empirical Characteristic Function displays good statistical properties in detecting misspecifications of different nature in the forecasts
    corecore