83 research outputs found
Where do statistical models come from? Revisiting the problem of specification
R. A. Fisher founded modern statistical inference in 1922 and identified its
fundamental problems to be: specification, estimation and distribution. Since
then the problem of statistical model specification has received scant
attention in the statistics literature. The paper traces the history of
statistical model specification, focusing primarily on pioneers like Fisher,
Neyman, and more recently Lehmann and Cox, and attempts a synthesis of their
views in the context of the Probabilistic Reduction (PR) approach. As argued by
Lehmann [11], a major stumbling block for a general approach to statistical
model specification has been the delineation of the appropriate role for
substantive subject matter information. The PR approach demarcates the
interrelated but complemenatry roles of substantive and statistical information
summarized ab initio in the form of a structural and a statistical model,
respectively. In an attempt to preserve the integrity of both sources of
information, as well as to ensure the reliability of their fusing, a purely
probabilistic construal of statistical models is advocated. This probabilistic
construal is then used to shed light on a number of issues relating to
specification, including the role of preliminary data analysis, structural vs.
statistical models, model specification vs. model selection, statistical vs.
substantive adequacy and model validation.Comment: Published at http://dx.doi.org/10.1214/074921706000000419 in the IMS
Lecture Notes--Monograph Series
(http://www.imstat.org/publications/lecnotes.htm) by the Institute of
Mathematical Statistics (http://www.imstat.org
Revisiting the Neyman-Scott model: an Inconsistent MLE or an Ill-defined Model?
The Neyman and Scott (1948) model is widely used to demonstrate a serious
weakness of the Maximum Likelihood (ML) method: it can give rise to
inconsistent estimators. The primary objective of this paper is to revisit this
example with a view to demonstrate that the culprit for the inconsistent
estimation is not the ML method but an ill-defined statistical model. It is
also shown that a simple recasting of this model renders it well-defined and
the ML method gives rise to consistent and asymptotically efficient estimators
Why the DecisionâTheoretic Perspective Misrepresents Frequentist Inference: Revisiting Steinâs Paradox and Admissibility
The primary objective of this paper is to make a case that R.A. Fisherâs objections to the decisionâtheoretic framing of frequentist inference are not without merit. It is argued that this framing is congruent with the Bayesian but incongruent with the frequentist approach; it provides the former with a theory of optimal inference but misrepresents the optimality theory of the latter. Decisionâtheoretic and Bayesian rules are considered optimal when they minimize the expected loss âfor all possible values of θ in Îâ [âθâÎ], irrespective of what the true value θâ [state of Nature] happens to be; the value that gave rise to the data. In contrast, the theory of optimal frequentist inference is framed entirely in terms of the capacity of the procedure to pinpoint θâ. The inappropriateness of the quantifier âθâÎ calls into question the relevance of admissibility as a minimal property for frequentist estimators. As a result, the pertinence of Steinâs paradox, as it relates to the capacity of frequentist estimators to pinpoint θâ, needs to be reassessed. The paper also contrasts lossâbased errors with traditional frequentist errors, arguing that the former are attached to θ, but the latter to the inference procedure itself
THE LINEAR REGRESSION MODEL WITH AUTOCORRELATED ERRORS: JUST SAY NO TO ERROR AUTOCORRELATION
This paper focuses on the practice of serial correlation correcting of the Linear Regression Model (LRM) by modeling the error. Simple Monte Carlo experiments are used to demonstrate the following points regarding this practice. First, the common factor restrictions implicitly imposed on the temporal structure of yt and xt appear to be completely unreasonable for any real world application. Second, when one compares the Autocorrelation-Corrected LRM (ACLRM) model estimates with estimates from the (unrestricted) Dynamic Linear Regression Model (DLRM) encompassing the ACLRM there is no significant gain in efficiency! Third, as expected, when the common factor restrictions do not hold the LRM model gives poor estimates of the true parameters and estimation of the ACLRM simply gives rise to different misleading results! On the other hand, estimates from the DLRM and the corresponding VAR model are very reliable. Fourth, the power of the usual Durbin Watson test (DW) of autocorrelation is much higher when the common factor restrictions do hold than when they do not. But, a more general test of autocorrelation is shown to perform almost as well as the DW when the common factor restrictions do hold and significantly better than the DW when the restrictions do not hold. Fifth, we demonstrate how simple it is to, at least, test the common factor restrictions imposed and we illustrate how powerful this test can be.Research Methods/ Statistical Methods,
Bernoulli Regression Models: Re-examining Statistical Models with Binary Dependent Variables
The classical approach for specifying statistical models with binary dependent variables in econometrics using latent variables or threshold models can leave the model misspecified, resulting in biased and inconsistent estimates as well as erroneous inferences. Furthermore, methods for trying to alleviate such problems, such as univariate generalized linear models, have not provided an adequate alternative for ensuring the statistical adequacy of such models. The purpose of this paper is to re-examine the underlying probabilistic foundations of statistical models with binary dependent variables using the probabilistic reduction approach to provide an alternative approach for model specification. This re-examination leads to the development of the Bernoulli Regression Model. Simulated and empirical examples provide evidence that the Bernoulli Regression Model can provide a superior approach for specifying statistically adequate models for dichotomous choice processes.Bernoulli Regression Model, logistic regression, generalized linear models, discrete choice, probabilistic reduction approach, model specification, Research Methods/ Statistical Methods,
Testing for Structural Breaks and other forms of Non-stationarity: a Misspecification Perspective
In the 1980s and 1990s the issue of non-stationarity in economic time series has been in the context of unit roots vs. mean trends in AR(p) models. More recently this perspective has been extended to include structural breaks. In this paper we take a much broader perspective by viewing the problem as one of misspecification testing: assessing the stationarity of the underlying process. The proposed misspecification testing procedure relies on resampling techniques to enhance the informational content of the observed data in an attempt to capture heterogeneity `locally' using rolling window estimators of the primary moments of the stochastic process. The effectiveness of the testing procedure is assessed using extensive Monte Carlo simulationsMaximum Entropy Bootstrap, Non-Stationarity
What Foundations for Statistical Modeling and Inference?
The primary aim of this article is to review the above books in a comparative way from the standpoint of my perspective on empirical modeling and inference. 1 Hacking (1965). Logic of Statistical Inference 2. Mayo (2018). Statistical Inference as Severe Testing: How to Get Beyond the Statistics Wars 3. Conclusion
Bernoulli Regression Models: Revisiting the Specification of Statistical Models with Binary Dependent Variables
The latent variable and generalized linear modelling approaches do not provide a systematic approach for modelling discrete choice observational data. Another alternative, the probabilistic reduction (PR) approach, provides a systematic way to specify such models that can yield reliable statistical and substantive inferences. The purpose of this paper is to re-examine the underlying probabilistic foundations of conditional statistical models with binary dependent variables using the PR approach. This leads to the development of the Bernoulli Regression Model, a family of statistical models, which includes the binary logistic regression model. The paper provides an explicit presentation of probabilistic model assumptions, guidance on model specification and estimation, and empirical application
- âŚ