6,032 research outputs found
Wage Mobility in East and West Germany
This article studies the long run patterns and explanations of wage mobility as a characteristic of regional labor markets. Using German administrative data we describe wage mobility since 1975 in West and since 1992 in East Germany. Wage mobility declined substantially in East Germany in the 1990s and moderately in East and West Germany since the late 1990s. Therefore, wage mobility does not balance recent increases in cross-sectional wage inequality. We apply RIF (recentered influence function) regression based decompositions to measure the role of potential explanatory factors behind these mobility changes. Increasing job stability is an important factor associated with the East German mobility decline.wage mobility, earnings mobility, income mobility, Germany, East Germany, inequality, transition matrix, Shorrocks index, administrative data
Drifting Together or Falling Apart? The Empirics of Regional Economic Growth in Post-Unification Germany
The objective of this paper is to address the question of convergence across German districts in the first decade after German unification by drawing out and emphasising some stylised facts of regional per capita income dynamics. We achieve this by employing non-parametric techniques which focus on the evolution of the entire cross-sectional income distribution. In particular, we follow a distributional approach to convergence based on kernel density estimation and implement a number of tests to establish the statistical significance of our findings. This paper finds that the relative income distribution appears to be stratifying into a trimodal/bimodal distribution.regional economic growth, Germany, convergence clubs, density estimation, modality tests
Bayesian Cognitive Science, Unification, and Explanation
It is often claimed that the greatest value of the Bayesian framework in cognitive science consists in its unifying power. Several Bayesian cognitive scientists assume that unification is obviously linked to explanatory power. But this link is not obvious, as unification in science is a heterogeneous notion, which may have little to do with explanation. While a crucial feature of most adequate explanations in cognitive science is that they reveal aspects of the causal mechanism that produces the phenomenon to be explained, the kind of unification afforded by the Bayesian framework to cognitive science does not necessarily reveal aspects of a mechanism. Bayesian unification, nonetheless, can place fruitful constraints on causal-mechanical explanation
Getting started in probabilistic graphical models
Probabilistic graphical models (PGMs) have become a popular tool for
computational analysis of biological data in a variety of domains. But, what
exactly are they and how do they work? How can we use PGMs to discover patterns
that are biologically relevant? And to what extent can PGMs help us formulate
new hypotheses that are testable at the bench? This note sketches out some
answers and illustrates the main ideas behind the statistical approach to
biological pattern discovery.Comment: 12 pages, 1 figur
Flexible parametric bootstrap for testing homogeneity against clustering and assessing the number of clusters
There are two notoriously hard problems in cluster analysis, estimating the
number of clusters, and checking whether the population to be clustered is not
actually homogeneous. Given a dataset, a clustering method and a cluster
validation index, this paper proposes to set up null models that capture
structural features of the data that cannot be interpreted as indicating
clustering. Artificial datasets are sampled from the null model with parameters
estimated from the original dataset. This can be used for testing the null
hypothesis of a homogeneous population against a clustering alternative. It can
also be used to calibrate the validation index for estimating the number of
clusters, by taking into account the expected distribution of the index under
the null model for any given number of clusters. The approach is illustrated by
three examples, involving various different clustering techniques (partitioning
around medoids, hierarchical methods, a Gaussian mixture model), validation
indexes (average silhouette width, prediction strength and BIC), and issues
such as mixed type data, temporal and spatial autocorrelation
Bayesian Cognitive Science, Unification, and Explanation
It is often claimed that the greatest value of the Bayesian framework in cognitive science consists in its unifying power. Several Bayesian cognitive scientists assume that unification is obviously linked to explanatory power. But this link is not obvious, as unification in science is a heterogeneous notion, which may have little to do with explanation. While a crucial feature of most adequate explanations in cognitive science is that they reveal aspects of the causal mechanism that produces the phenomenon to be explained, the kind of unification afforded by the Bayesian framework to cognitive science does not necessarily reveal aspects of a mechanism. Bayesian unification, nonetheless, can place fruitful constraints on causal–mechanical explanation. 1 Introduction2 What a Great Many Phenomena Bayesian Decision Theory Can Model3 The Case of Information Integration4 How Do Bayesian Models Unify?5 Bayesian Unification: What Constraints Are There on Mechanistic Explanation?5.1 Unification constrains mechanism discovery5.2 Unification constrains the identification of relevant mechanistic factors5.3 Unification constrains confirmation of competitive mechanistic models6 ConclusionAppendix
Sample size and power estimation for studies with health related quality of life outcomes: a comparison of four methods using the SF-36
We describe and compare four different methods for estimating sample size and power, when the primary outcome of the study is a Health Related Quality of Life (HRQoL) measure. These methods are: 1. assuming a Normal distribution and comparing two means; 2. using a non-parametric method; 3. Whitehead's method based on the proportional odds model; 4. the bootstrap. We illustrate the various methods, using data from the SF-36. For simplicity this paper deals with studies designed to compare the effectiveness (or superiority) of a new treatment compared to a standard treatment at a single point in time. The results show that if the HRQoL outcome has a limited number of discrete values (< 7) and/or the expected proportion of cases at the boundaries is high (scoring 0 or 100), then we would recommend using Whitehead's method (Method 3). Alternatively, if the HRQoL outcome has a large number of distinct values and the proportion at the boundaries is low, then we would recommend using Method 1. If a pilot or historical dataset is readily available (to estimate the shape of the distribution) then bootstrap simulation (Method 4) based on this data will provide a more accurate and reliable sample size estimate than conventional methods (Methods 1, 2, or 3). In the absence of a reliable pilot set, bootstrapping is not appropriate and conventional methods of sample size estimation or simulation will need to be used. Fortunately, with the increasing use of HRQoL outcomes in research, historical datasets are becoming more readily available. Strictly speaking, our results and conclusions only apply to the SF-36 outcome measure. Further empirical work is required to see whether these results hold true for other HRQoL outcomes. However, the SF-36 has many features in common with other HRQoL outcomes: multi-dimensional, ordinal or discrete response categories with upper and lower bounds, and skewed distributions, so therefore, we believe these results and conclusions using the SF-36 will be appropriate for other HRQoL measures
Recommended from our members
Econometrics: A bird's eye view
As a unified discipline, econometrics is still relatively young and has been transforming and expanding very rapidly over the past few decades. Major advances have taken place in the analysis of cross sectional data by means of semi-parametric and non-parametric techniques. Heterogeneity of economic relations across individuals, firms and industries is increasingly acknowledge and attempts have been made to take them into account either by integrating out their effects or by modeling the sources of heterogeneity when suitable panel data exists. The counterfactual considerations that underlie policy analysis and treatment evaluation have been given a more satisfactory foundation. New time series econometric techniques have been developed and employed extensively in the areas of macroeconometrics and finance. Non-linear econometric techniques are used increasingly in the analysis of cross section and time series observations. Applications of Bayesian techniques to econometric problems have been given new impetus largely thanks to advances in computer power and computational techniques. The use of Bayesian techniques have in turn provided the investigators with a unifying framework where the tasks and forecasting, decision making, model evaluation and learning can be considered as parts of the same interactive and iterative process; thus paving the way for establishing the foundation of the "real time econometrics". This paper attempts to provide an overview of some of these developments
- …