100 research outputs found
The Jeffreys-Lindley Paradox and Discovery Criteria in High Energy Physics
The Jeffreys-Lindley paradox displays how the use of a p-value (or number of
standard deviations z) in a frequentist hypothesis test can lead to an
inference that is radically different from that of a Bayesian hypothesis test
in the form advocated by Harold Jeffreys in the 1930s and common today. The
setting is the test of a well-specified null hypothesis (such as the Standard
Model of elementary particle physics, possibly with "nuisance parameters")
versus a composite alternative (such as the Standard Model plus a new force of
nature of unknown strength). The p-value, as well as the ratio of the
likelihood under the null hypothesis to the maximized likelihood under the
alternative, can strongly disfavor the null hypothesis, while the Bayesian
posterior probability for the null hypothesis can be arbitrarily large. The
academic statistics literature contains many impassioned comments on this
paradox, yet there is no consensus either on its relevance to scientific
communication or on its correct resolution. The paradox is quite relevant to
frontier research in high energy physics. This paper is an attempt to explain
the situation to both physicists and statisticians, in the hope that further
progress can be made.Comment: v4: Continued editing for clarity. Figure added. v5: Minor fixes to
biblio. Same as published version except for minor copy-edits, Synthese
(2014). v6: fix typos, and restore garbled sentence at beginning of Sec 4 to
v
Pragmatic Warrant for Frequentist Statistical Practice: The Case of High Energy Physics
Amidst long-standing debates within the field, High Energy Physics (HEP) has adopted a statistical methodology that employs primarily standard frequentist techniques such as significance testing and confidence interval estimation, but incorporates Bayesian methods for limited purposes. The recent discovery of the Higgs boson has drawn increased attention to the statistical methods employed within HEP, leading to reprisals of numerous well-worn arguments amongst frequentists and Bayesians. Here I argue that the warrant for the practice in HEP of relying primarily on frequentist methods can best be understood within the tradition of philosophical pragmatism. In particular, I argue that understanding the statistical methodology of HEP through the perspective of pragmatism clarifies the role of and rationale for significance testing in the search for new phenomena such as the Higgs boson
Limitations of Bayesian Evidence Applied to Cosmology
There has been increasing interest by cosmologists in applying Bayesian
techniques, such as Bayesian Evidence, for model selection. A typical example
is in assessing whether observational data favour a cosmological constant over
evolving dark energy. In this paper, the example of dark energy is used to
illustrate limitations in the application of Bayesian Evidence associated with
subjective judgements concerning the choice of model and priors. An analysis of
recent cosmological data shows a statistically insignificant preference for a
cosmological constant over simple dynamical models of dark energy. It is argued
that for nested problems, as considered here, Bayesian parameter estimation can
be more informative than computing Bayesian Evidence for poorly motivated
physical models.Comment: 8 pages 4 figures MNRAS accepted. Substantially revised and extende
A Philosophical Look at the Discovery of the Higgs Boson
An Introduction to the Special Issue of the same name that appeared in Synthese 194/2, 201
Testing Point Null Hypothesis of a Normal Mean and the Truth: 21st Century Perspective
Testing a point (sharp) null hypothesis is arguably the most widely used statistical inferential procedure in many fields of scientific research, nevertheless, the most controversial, and misapprehended. Since 1935 when Buchanan-Wollaston raised the first criticism against hypothesis testing, this foundational field of statistics has drawn increasingly active and stronger opposition, including draconian suggestions that statistical significance testing should be abandoned or even banned. Statisticians should stop ignoring these accumulated and significant anomalies within the current point-null hypotheses paradigm and rebuild healthy foundations of statistical science. The foundation for a paradigm shift in testing statistical hypotheses is suggested, which is testing interval null hypotheses based on implications of the Zero probability paradox. It states that in a real-world research point-null hypothesis of a normal mean has zero probability. This implies that formulated point-null hypothesis of a mean in the context of the simple normal model is almost surely false. Thus, Zero probability paradox points to the root cause of so-called large n problem in significance testing. It discloses that there is no point in searching for a cure under the current point-null paradigm
Statistical Issues in Searches for New Physics
Given the cost, both financial and even more importantly in terms of human
effort, in building High Energy Physics accelerators and detectors and running
them, it is important to use good statistical techniques in analysing data.
Some of the statistical issues that arise in searches for New Physics are
discussed briefly. They include topics such as:
Should we insist on the 5 sigma criterion for discovery claims?
The probability of A, given B, is not the same as the probability of B, given
A.
The meaning of p-values.
What is Wilks Theorem and when does it not apply?
How should we deal with the `Look Elsewhere Effect'?
Dealing with systematics such as background parametrisation.
Coverage: What is it and does my method have the correct coverage?
The use of p0 versus p1 plots.Comment: This is the write-up for the Proceedings of a talk delivered at the
LHCP2014 Conference at Columbia University, New York in June 2014no diagrams
5 pages long
On the Authentic Notion, Relevance, and Solution of the Jeffreys-Lindley Paradox in the Zettabyte Era
The Jeffreys-Lindley paradox is the most quoted divergence between the frequentist and Bayesian approaches to statistical inference. It is embedded in the very foundations of statistics and divides frequentist and Bayesian inference in an irreconcilable way. This paradox is the Gordian Knot of statistical inference and Data Science in the Zettabyte Era. If statistical science is ready for revolution confronted by the challenges of massive data sets analysis, the first step is to finally solve this anomaly. For more than sixty years, the Jeffreys-Lindley paradox has been under active discussion and debate. Many solutions have been proposed, none entirely satisfactory. The Jeffreys-Lindley paradox and its extent have been frequently misunderstood by many statisticians and non-statisticians. This paper aims to reassess this paradox, shed new light on it, and indicates how often it occurs in practice when dealing with Big data
On the Authentic Notion, Relevance, and Solution of the Jeffreys-Lindley Paradox in the Zettabyte Era
The Jeffreys-Lindley paradox is the most quoted divergence between the frequentist and Bayesian approaches to statistical inference. It is embedded in the very foundations of statistics and divides frequentist and Bayesian inference in an irreconcilable way. This paradox is the Gordian Knot of statistical inference and Data Science in the Zettabyte Era. If statistical science is ready for revolution confronted by the challenges of massive data sets analysis, the first step is to finally solve this anomaly. For more than sixty years, the Jeffreys-Lindley paradox has been under active discussion and debate. Many solutions have been proposed, none entirely satisfactory. The Jeffreys-Lindley paradox and its extent have been frequently misunderstood by many statisticians and non-statisticians. This paper aims to reassess this paradox, shed new light on it, and indicates how often it occurs in practice when dealing with Big data
- …