164 research outputs found
Ecological non-linear state space model selection via adaptive particle Markov chain Monte Carlo (AdPMCMC)
We develop a novel advanced Particle Markov chain Monte Carlo algorithm that
is capable of sampling from the posterior distribution of non-linear state
space models for both the unobserved latent states and the unknown model
parameters. We apply this novel methodology to five population growth models,
including models with strong and weak Allee effects, and test if it can
efficiently sample from the complex likelihood surface that is often associated
with these models. Utilising real and also synthetically generated data sets we
examine the extent to which observation noise and process error may frustrate
efforts to choose between these models. Our novel algorithm involves an
Adaptive Metropolis proposal combined with an SIR Particle MCMC algorithm
(AdPMCMC). We show that the AdPMCMC algorithm samples complex, high-dimensional
spaces efficiently, and is therefore superior to standard Gibbs or Metropolis
Hastings algorithms that are known to converge very slowly when applied to the
non-linear state space ecological models considered in this paper.
Additionally, we show how the AdPMCMC algorithm can be used to recursively
estimate the Bayesian Cram\'er-Rao Lower Bound of Tichavsk\'y (1998). We derive
expressions for these Cram\'er-Rao Bounds and estimate them for the models
considered. Our results demonstrate a number of important features of common
population growth models, most notably their multi-modal posterior surfaces and
dependence between the static and dynamic parameters. We conclude by sampling
from the posterior distribution of each of the models, and use Bayes factors to
highlight how observation noise significantly diminishes our ability to select
among some of the models, particularly those that are designed to reproduce an
Allee effect
Advances in approximate Bayesian computation and trans-dimensional sampling methodology
Bayesian statistical models continue to grow in complexity, driven
in part by a few key factors: the massive computational resources
now available to statisticians; the substantial gains made in
sampling methodology and algorithms such as Markov chain
Monte Carlo (MCMC), trans-dimensional MCMC (TDMCMC), sequential
Monte Carlo (SMC), adaptive algorithms and stochastic
approximation methods and approximate Bayesian computation (ABC);
and development of more realistic models for real world phenomena
as demonstrated in this thesis for financial models and
telecommunications engineering. Sophisticated statistical models
are increasingly proposed for practical solutions to real world problems in order to better capture salient features of
increasingly more complex data. With sophistication comes a
parallel requirement for more advanced and automated statistical
computational methodologies.
The key focus of this thesis revolves around innovation related to
the following three significant Bayesian research questions.
1. How can one develop practically useful Bayesian models and corresponding computationally efficient sampling methodology, when the likelihood model is intractable?
2. How can one develop methodology in order to automate Markov chain Monte Carlo sampling approaches to efficiently explore the support of a posterior distribution, defined across multiple Bayesian statistical models?
3. How can these sophisticated Bayesian modelling frameworks and sampling methodologies be utilized to solve practically relevant and important problems in the research fields of financial risk modeling and telecommunications engineering ?
This thesis is split into three bodies of work represented in
three parts. Each part contains journal papers with novel
statistical model and sampling methodological development. The
coherent link between each part involves the novel
sampling methodologies developed in Part I and utilized in Part II and Part III. Papers contained in
each part make progress at addressing the core research
questions posed.
Part I of this thesis presents generally applicable key
statistical sampling methodologies that will be utilized and
extended in the subsequent two parts. In particular it presents
novel developments in statistical methodology pertaining to
likelihood-free or ABC and TDMCMC methodology.
The TDMCMC methodology focuses on several aspects of automation
in the between model proposal construction, including
approximation of the optimal between model proposal kernel via a
conditional path sampling density estimator. Then this methodology
is explored for several novel Bayesian model selection
applications including cointegrated vector autoregressions (CVAR)
models and mixture models in which there is an unknown number of
mixture components. The second area relates to development of
ABC methodology with particular focus
on SMC Samplers methodology in an ABC context via Partial
Rejection Control (PRC). In addition to novel algorithmic
development, key theoretical properties are also studied for the
classes of algorithms developed. Then this methodology is
developed for a highly challenging practically significant
application relating to multivariate Bayesian -stable
models.
Then Part II focuses on novel statistical model development
in the areas of financial risk and non-life insurance claims
reserving. In each of the papers in this part the focus is on
two aspects: foremost the development of novel statistical models
to improve the modeling of risk and insurance; and then the
associated problem of how to fit and sample from such statistical
models efficiently. In particular novel statistical models are
developed for Operational Risk (OpRisk) under a Loss Distributional
Approach (LDA) and for claims reserving in Actuarial non-life
insurance modelling. In each case the models developed include an
additional level of complexity which adds flexibility to the model
in order to better capture salient features observed in real data.
The consequence of the additional complexity comes at the cost
that standard fitting and sampling methodologies are generally not
applicable, as a result one is required to develop and apply the
methodology from Part I.
Part III focuses on novel statistical model development
in the area of statistical signal processing for wireless
communications engineering. Statistical models will be developed
or extended for two general classes of wireless communications
problem: the first relates to detection of transmitted symbols and
joint channel estimation in Multiple Input Multiple Output (MIMO)
systems coupled with Orthogonal Frequency Division Multiplexing
(OFDM); the second relates to co-operative wireless communications
relay systems in which the key focus is on detection of
transmitted symbols. Both these areas will require advanced
sampling methodology developed in Part I to find solutions to
these real world engineering problems
Particle MCMC algorithms and architectures for accelerating inference in state-space models.
Particle Markov Chain Monte Carlo (pMCMC) is a stochastic algorithm designed to generate samples from a probability distribution, when the density of the distribution does not admit a closed form expression. pMCMC is most commonly used to sample from the Bayesian posterior distribution in State-Space Models (SSMs), a class of probabilistic models used in numerous scientific applications. Nevertheless, this task is prohibitive when dealing with complex SSMs with massive data, due to the high computational cost of pMCMC and its poor performance when the posterior exhibits multi-modality. This paper aims to address both issues by: 1) Proposing a novel pMCMC algorithm (denoted ppMCMC), which uses multiple Markov chains (instead of the one used by pMCMC) to improve sampling efficiency for multi-modal posteriors, 2) Introducing custom, parallel hardware architectures, which are tailored for pMCMC and ppMCMC. The architectures are implemented on Field Programmable Gate Arrays (FPGAs), a type of hardware accelerator with massive parallelization capabilities. The new algorithm and the two FPGA architectures are evaluated using a large-scale case study from genetics. Results indicate that ppMCMC achieves 1.96x higher sampling efficiency than pMCMC when using sequential CPU implementations. The FPGA architecture of pMCMC is 12.1x and 10.1x faster than state-of-the-art, parallel CPU and GPU implementations of pMCMC and up to 53x more energy efficient; the FPGA architecture of ppMCMC increases these speedups to 34.9x and 41.8x respectively and is 173x more power efficient, bringing previously intractable SSM-based data analyses within reach.The authors would like to thank the Wellcome Trust (Grant reference 097816/Z/11/A) and the EPSRC (Grant reference EP/I012036/1) for the financial support given to this research project
Predictive smart relaying schemes for decentralized wireless systems
Recent developments in decentralized wireless networks make the technology potentially deployable in an extremely broad scenarios and applications. These include mobile Internet of Things (IoT) networks, smart cities, future innovative communication systems with multiple aerial layer flying network platforms and other advanced mobile communication networks. The approach also could be the solution for traditional operated mobile network backup plans, balancing traffic flow, emergency communication systems and so on.
This thesis reveals and addresses several issues and challenges in conventional wireless communication systems, particular for the cases where there is a lack of resources and the disconnection of radio links. There are two message routing plans in the data packet store, carry and forwarding form are proposed, known as KaFiR and PaFiR. These employ the Bayesian filtering approach to track and predict the motion of surrounding portable devices and determine the next layer among candidate nodes. The relaying strategies endow smart devices with the intelligent capability to optimize the message routing path and improve the overall network performance with respect to resilience, tolerance and scalability.
The simulation and test results present that the KaFiR routing protocol performs well when network subscribers are less mobile and the relaying protocol can be deployed on a wide range of portable terminals as the algorithm is rather simple to operate. The PaFiR routing strategy takes advantages of the Particle Filter algorithm, which can cope with complex network scenarios and applications, particularly when unmanned aerial vehicles are involved as the assisted intermediate layers.
When compared with other existing DTN routing protocols and some of the latest relaying plans, both relaying protocols deliver an excellent overall performance for the key wireless communication network evolution metrics, which shows the promising future for this brand new research direction. Further extension work directions based on the tracking and prediction methods are suggested and reviewed. Future work on some new applications and services are also addressed
Cooperative Vehicle Tracking in Large Environments
Vehicle position tracking and prediction over large areas is of significant importance in many industrial applications, such as mining operations. In a small area, this can be easily achieved by providing vehicles with a constant communication link to a control centre and having the vehicles broadcast their position. The problem changes dramatically when vehicles operate within a large environment of potentially hundreds of square kilometres and in difficult terrain. This thesis presents algorithms for cooperative tracking of vehicles based on a vehicle motion model that incorporates the properties of the working area, and information collected by infrastructure collection points and other mobile agents. The probabilistic motion prediction approach provides long-term estimates of vehicle positions using motion profiles built for the particular environment and considering the vehicle stopping probability. A limited number of data collection points distributed around the field are used to update the position estimates, with negative information also used to improve the estimation. The thesis introduces the concept of observation harvesting, a process in which peer-to-peer communication between vehicles allows egocentric position updates and inter-vehicle measurements to be relayed among vehicles and finally conveyed to the collection points for an improved position estimate. It uses a store-and-synchronise concept to deal with intermittent communication and aims to disseminate data in an opportunistic manner. A nonparametric filtering algorithm for cooperative tracking is proposed to incorporate the information harvested, including the negative, relative, and time delayed observations. An important contribution of this thesis is to enable the optimisation of fleet scheduling when full coverage networks are not available or feasible. The proposed approaches were validated with comprehensive experimental results using data collected from a large-scale mining operation
Cooperative Vehicle Tracking in Large Environments
Vehicle position tracking and prediction over large areas is of significant importance in many industrial applications, such as mining operations. In a small area, this can be easily achieved by providing vehicles with a constant communication link to a control centre and having the vehicles broadcast their position. The problem changes dramatically when vehicles operate within a large environment of potentially hundreds of square kilometres and in difficult terrain. This thesis presents algorithms for cooperative tracking of vehicles based on a vehicle motion model that incorporates the properties of the working area, and information collected by infrastructure collection points and other mobile agents. The probabilistic motion prediction approach provides long-term estimates of vehicle positions using motion profiles built for the particular environment and considering the vehicle stopping probability. A limited number of data collection points distributed around the field are used to update the position estimates, with negative information also used to improve the estimation. The thesis introduces the concept of observation harvesting, a process in which peer-to-peer communication between vehicles allows egocentric position updates and inter-vehicle measurements to be relayed among vehicles and finally conveyed to the collection points for an improved position estimate. It uses a store-and-synchronise concept to deal with intermittent communication and aims to disseminate data in an opportunistic manner. A nonparametric filtering algorithm for cooperative tracking is proposed to incorporate the information harvested, including the negative, relative, and time delayed observations. An important contribution of this thesis is to enable the optimisation of fleet scheduling when full coverage networks are not available or feasible. The proposed approaches were validated with comprehensive experimental results using data collected from a large-scale mining operation
Dynamics of Information Diffusion and Social Sensing
Statistical inference using social sensors is an area that has witnessed
remarkable progress and is relevant in applications including localizing events
for targeted advertising, marketing, localization of natural disasters and
predicting sentiment of investors in financial markets. This chapter presents a
tutorial description of four important aspects of sensing-based information
diffusion in social networks from a communications/signal processing
perspective. First, diffusion models for information exchange in large scale
social networks together with social sensing via social media networks such as
Twitter is considered. Second, Bayesian social learning models and risk averse
social learning is considered with applications in finance and online
reputation systems. Third, the principle of revealed preferences arising in
micro-economics theory is used to parse datasets to determine if social sensors
are utility maximizers and then determine their utility functions. Finally, the
interaction of social sensors with YouTube channel owners is studied using time
series analysis methods. All four topics are explained in the context of actual
experimental datasets from health networks, social media and psychological
experiments. Also, algorithms are given that exploit the above models to infer
underlying events based on social sensing. The overview, insights, models and
algorithms presented in this chapter stem from recent developments in network
science, economics and signal processing. At a deeper level, this chapter
considers mean field dynamics of networks, risk averse Bayesian social learning
filtering and quickest change detection, data incest in decision making over a
directed acyclic graph of social sensors, inverse optimization problems for
utility function estimation (revealed preferences) and statistical modeling of
interacting social sensors in YouTube social networks.Comment: arXiv admin note: text overlap with arXiv:1405.112
ADAPTIVE CHANNEL AND SOURCE CODING USING APPROXIMATE INFERENCE
Channel coding and source coding are two important problems in communications. Although both channel coding and source coding (especially, the distributed source coding (DSC)) can achieve their ultimate performance by knowing the perfect knowledge of channel noise and source correlation, respectively, such information may not be always available at the decoder side. The reasons might be because of the time−varying characteristic of some communication systems and sources themselves, respectively. In this dissertation, I mainly focus on the study of online channel noise estimation and correlation estimation by using both stochastic and deterministic approximation inferences on factor graphs.In channel coding, belief propagation (BP) is a powerful algorithm to decode low−density parity check (LDPC) codes over additive white Gaussian noise (AWGN) channels. However, the traditional BP algorithm cannot adapt efficiently to the statistical change of SNR in an AWGN channel. To solve the problem, two common workarounds in approximate inference are stochastic methods (e.g. particle filtering (PF)) and deterministic methods (e.g. expectation approximation (EP)). Generally, deterministic methods are much faster than stochastic methods. In contrast, stochastic methods are more flexible and suitable for any distribution. In this dissertation, I proposed two adaptive LDPC decoding schemes, which are able to perform online estimation of time−varying channel state information (especially signal to noise ratio (SNR)) at the bit−level by incorporating PF and EP algorithms. Through experimental results, I compare the performance between the proposed PF based and EP based approaches, which shows that the EP based approach obtains the comparable estimation accuracy with less computational complexity than the PF based method for both stationary and time−varying SNR, and enhances the BP decoding performance simultaneously. Moreover, the EP estimator shows a very fast convergence speed, and the additional computational overhead of the proposed decoder is less than 10% of the standard BP decoder.Moreover, since the close relationship between source coding and channel coding, the proposed ideas are extended to source correlation estimation. First, I study the correlation estimation problem in lossless DSC setup, where I consider both asymmetric and non−asymmetric SW coding of two binary correlated sources. The aforementioned PF and EP based approaches are extended to handle the correlation between two binary sources, where the relationship is modeled as a virtual binary symmetric channel (BSC) with a time−varying crossover probability. Besides, to handle the correlation estimation problem of Wyner−Ziv (WZ) coding, a lossy DSC setup, I design a joint bit−plane model, by which the PF based approach can be applied to tracking the correlation between non−binary sources. Through experimental results, the proposed correlation estimation approaches significantly improve the compression performance of DSC.Finally, due to the property of ultra−low encoding complexity, DSC is a promising technique for many tasks, in which the encoder has only limited computing and communication power, e.g. the space imaging systems. In this dissertation, I consider a real−world application of the proposed correlation estimation scheme on the onboard low−complexity compression of solar stereo images, since such solutions are essential to reduce onboard storage, processing, and communication resources. In this dissertation, I propose an adaptive distributed compression solution using PF that tracks the correlation, as well as performs disparity estimation, at the decoder side. The proposed algorithm istested on the stereo solar images captured by the twin satellites systemof NASA’s STEREO project. The experimental results show the significant PSNR improvement over traditional separate bit−plane decoding without dynamic correlation and disparity estimation
- …