446 research outputs found
Switching Regression Models and Causal Inference in the Presence of Discrete Latent Variables
Given a response and a vector of predictors,
we investigate the problem of inferring direct causes of among the vector
. Models for that use all of its causal covariates as predictors enjoy
the property of being invariant across different environments or interventional
settings. Given data from such environments, this property has been exploited
for causal discovery. Here, we extend this inference principle to situations in
which some (discrete-valued) direct causes of are unobserved. Such cases
naturally give rise to switching regression models. We provide sufficient
conditions for the existence, consistency and asymptotic normality of the MLE
in linear switching regression models with Gaussian noise, and construct a test
for the equality of such models. These results allow us to prove that the
proposed causal discovery method obtains asymptotic false discovery control
under mild conditions. We provide an algorithm, make available code, and test
our method on simulated data. It is robust against model violations and
outperforms state-of-the-art approaches. We further apply our method to a real
data set, where we show that it does not only output causal predictors, but
also a process-based clustering of data points, which could be of additional
interest to practitioners.Comment: 46 pages, 14 figures; real-world application added in Section 5.2;
additional numerical experiments added in the Appendix
Identifiability of generalised Randles circuit models
The Randles circuit (including a parallel resistor and capacitor in series
with another resistor) and its generalised topology have widely been employed
in electrochemical energy storage systems such as batteries, fuel cells and
supercapacitors, also in biomedical engineering, for example, to model the
electrode-tissue interface in electroencephalography and baroreceptor dynamics.
This paper studies identifiability of generalised Randles circuit models, that
is, whether the model parameters can be estimated uniquely from the
input-output data. It is shown that generalised Randles circuit models are
structurally locally identifiable. The condition that makes the model structure
globally identifiable is then discussed. Finally, the estimation accuracy is
evaluated through extensive simulations
Multilevel linear models, Gibbs samplers and multigrid decompositions (with Discussion)
We study the convergence properties of the Gibbs Sampler in the context of posterior distributions arising from Bayesian analysis of conditionally Gaussian hierarchical models. We develop a multigrid approach to derive analytic expressions for the convergence rates of the algorithm for various widely used model structures, including nested and crossed random effects. Our results apply to multilevel models with an arbitrary number of layers in the hierarchy, while most previous work was limited to the two-level nested case. The theoretical results provide explicit and easy-to-implement guidelines to optimize practical implementations of the Gibbs Sampler, such as indications on which parametrization to choose (e.g. centred and non-centred), which constraint to impose to guarantee statistical identifiability, and which parameters to monitor in the diagnostic process. Simulations suggest that the results are informative also in the context of non-Gaussian distributions and more general MCMC schemes, such as gradient-based ones
Dynamical compensation and structural identifiability: analysis, implications, and reconciliation
The concept of dynamical compensation has been recently introduced to
describe the ability of a biological system to keep its output dynamics
unchanged in the face of varying parameters. Here we show that, according to
its original definition, dynamical compensation is equivalent to lack of
structural identifiability. This is relevant if model parameters need to be
estimated, which is often the case in biological modelling. This realization
prompts us to warn that care should we taken when using an unidentifiable model
to extract biological insight: the estimated values of structurally
unidentifiable parameters are meaningless, and model predictions about
unmeasured state variables can be wrong. Taking this into account, we explore
alternative definitions of dynamical compensation that do not necessarily imply
structural unidentifiability. Accordingly, we show different ways in which a
model can be made identifiable while exhibiting dynamical compensation. Our
analyses enable the use of the new concept of dynamical compensation in the
context of parameter identification, and reconcile it with the desirable
property of structural identifiability
Observability and Structural Identifiability of Nonlinear Biological Systems
Observability is a modelling property that describes the possibility of
inferring the internal state of a system from observations of its output. A
related property, structural identifiability, refers to the theoretical
possibility of determining the parameter values from the output. In fact,
structural identifiability becomes a particular case of observability if the
parameters are considered as constant state variables. It is possible to
simultaneously analyse the observability and structural identifiability of a
model using the conceptual tools of differential geometry. Many complex
biological processes can be described by systems of nonlinear ordinary
differential equations, and can therefore be analysed with this approach. The
purpose of this review article is threefold: (I) to serve as a tutorial on
observability and structural identifiability of nonlinear systems, using the
differential geometry approach for their analysis; (II) to review recent
advances in the field; and (III) to identify open problems and suggest new
avenues for research in this area.Comment: Accepted for publication in the special issue "Computational Methods
for Identification and Modelling of Complex Biological Systems" of Complexit
Error probability analysis in quantum tomography: a tool for evaluating experiments
We expand the scope of the statistical notion of error probability, i.e., how
often large deviations are observed in an experiment, in order to make it
directly applicable to quantum tomography. We verify that the error probability
can decrease at most exponentially in the number of trials, derive the explicit
rate that bounds this decrease, and show that a maximum likelihood estimator
achieves this bound. We also show that the statistical notion of
identifiability coincides with the tomographic notion of informational
completeness. Our result implies that two quantum tomographic apparatuses that
have the same risk function, (e.g. variance), can have different error
probability, and we give an example in one qubit state tomography. Thus by
combining these two approaches we can evaluate, in a reconstruction independent
way, the performance of such experiments more discerningly.Comment: 14pages, 2 figures (an analysis of an example is added, and the proof
of Lemma 2 is corrected
- …