110 research outputs found
Recommended from our members
Approaches to Safety in Inverse Reinforcement Learning
As the capabilities of robotic systems increase, we move closer to the vision of ubiquitous robotic assistance throughout our everyday lives. In transitioning robots and autonomous systems from traditional factory and industrial settings, it is critical that these systems are able to adapt to uncertain environments and the humans who populate them. In order to better understand and predict the behavior of these humans, Inverse Reinforcement Learning (IRL) uses demonstrations to infer the underlying motivations driving human actions. The information gained from IRL can be used to improve a robot’s understanding of the environment as well as to allow the robot to better interact with or assist humans.In this dissertation, we address the challenge of incorporating safety into the application of IRL. We first consider safety in the context of using IRL for assisting humans in shared control tasks. Through a user study, we show how incorporating haptic feedback into human assistance can increase humans’ sense of control while improving safety in the presence of imperfect learning. Further, we present our method for using IRL to automatically create such haptic feedback policies from task demonstrations.We further address safety in IRL by incorporating notions of safety directly into the learning process. Currently, most work on IRL focuses on learning explanatory rewards that humans are modeled as optimizing. However, pure reward optimization can fail to effectively capture hard requirements, such as safety constraints. We draw on the definition of safety from Hamilton-Jacobi reachability analysis to infer human perceptions of safety and to modify robot behavior to respect these learned safety constraints. We also extend this work on learning constraints by adapting the framework of Maximum Entropy IRL in order to learn hard constraints given nominal task rewards, and we show how this technique infers the most likely constraints to align expected behavior with observed demonstrations
Advances in approximate Bayesian computation and trans-dimensional sampling methodology
Bayesian statistical models continue to grow in complexity, driven
in part by a few key factors: the massive computational resources
now available to statisticians; the substantial gains made in
sampling methodology and algorithms such as Markov chain
Monte Carlo (MCMC), trans-dimensional MCMC (TDMCMC), sequential
Monte Carlo (SMC), adaptive algorithms and stochastic
approximation methods and approximate Bayesian computation (ABC);
and development of more realistic models for real world phenomena
as demonstrated in this thesis for financial models and
telecommunications engineering. Sophisticated statistical models
are increasingly proposed for practical solutions to real world problems in order to better capture salient features of
increasingly more complex data. With sophistication comes a
parallel requirement for more advanced and automated statistical
computational methodologies.
The key focus of this thesis revolves around innovation related to
the following three significant Bayesian research questions.
1. How can one develop practically useful Bayesian models and corresponding computationally efficient sampling methodology, when the likelihood model is intractable?
2. How can one develop methodology in order to automate Markov chain Monte Carlo sampling approaches to efficiently explore the support of a posterior distribution, defined across multiple Bayesian statistical models?
3. How can these sophisticated Bayesian modelling frameworks and sampling methodologies be utilized to solve practically relevant and important problems in the research fields of financial risk modeling and telecommunications engineering ?
This thesis is split into three bodies of work represented in
three parts. Each part contains journal papers with novel
statistical model and sampling methodological development. The
coherent link between each part involves the novel
sampling methodologies developed in Part I and utilized in Part II and Part III. Papers contained in
each part make progress at addressing the core research
questions posed.
Part I of this thesis presents generally applicable key
statistical sampling methodologies that will be utilized and
extended in the subsequent two parts. In particular it presents
novel developments in statistical methodology pertaining to
likelihood-free or ABC and TDMCMC methodology.
The TDMCMC methodology focuses on several aspects of automation
in the between model proposal construction, including
approximation of the optimal between model proposal kernel via a
conditional path sampling density estimator. Then this methodology
is explored for several novel Bayesian model selection
applications including cointegrated vector autoregressions (CVAR)
models and mixture models in which there is an unknown number of
mixture components. The second area relates to development of
ABC methodology with particular focus
on SMC Samplers methodology in an ABC context via Partial
Rejection Control (PRC). In addition to novel algorithmic
development, key theoretical properties are also studied for the
classes of algorithms developed. Then this methodology is
developed for a highly challenging practically significant
application relating to multivariate Bayesian -stable
models.
Then Part II focuses on novel statistical model development
in the areas of financial risk and non-life insurance claims
reserving. In each of the papers in this part the focus is on
two aspects: foremost the development of novel statistical models
to improve the modeling of risk and insurance; and then the
associated problem of how to fit and sample from such statistical
models efficiently. In particular novel statistical models are
developed for Operational Risk (OpRisk) under a Loss Distributional
Approach (LDA) and for claims reserving in Actuarial non-life
insurance modelling. In each case the models developed include an
additional level of complexity which adds flexibility to the model
in order to better capture salient features observed in real data.
The consequence of the additional complexity comes at the cost
that standard fitting and sampling methodologies are generally not
applicable, as a result one is required to develop and apply the
methodology from Part I.
Part III focuses on novel statistical model development
in the area of statistical signal processing for wireless
communications engineering. Statistical models will be developed
or extended for two general classes of wireless communications
problem: the first relates to detection of transmitted symbols and
joint channel estimation in Multiple Input Multiple Output (MIMO)
systems coupled with Orthogonal Frequency Division Multiplexing
(OFDM); the second relates to co-operative wireless communications
relay systems in which the key focus is on detection of
transmitted symbols. Both these areas will require advanced
sampling methodology developed in Part I to find solutions to
these real world engineering problems
Modelling of Floods in Urban Areas
This Special Issue publishes the latest advances and developments concerning the modelling of flooding in urban areas and contributes to our scientific understanding of the flooding processes and the appropriate evaluation of flood impacts. This issue contains contributions of novel methodologies including flood forecasting methods, data acquisition techniques, experimental research in urban drainage systems and/or sustainable drainage systems, and new numerical and simulation approaches in nine papers with contributions from over forty authors
Stability and inference in discrete diffusion scale-spaces
Taking averages of observations is the most basic method to make inferences in the presence of uncertainty. In late 1980's, this simple idea has been extended to the principle of successively average less where the change is faster, and applied to the problem of revealing a signal with jump discontinuities in additive noise.
Successive averaging results in a family of signals with progressively decreasing amount of details, which is called the scale-space and further conveniently formalized by viewing it as a solution to a certain diffusion-inspired evolutionary partial differential equation (PDE). Such a model is known as the diffusion scale-space and it possesses two long-standing problems: (i) model analysis which aims at establishing stability and guarantees that averaging does not distort important information, and (ii) model selection, such as identification of the optimal scale (diffusion stopping time) given an initial noisy signal and an incomplete model.
This thesis studies both problems in the discrete space and time. Such a setting has been strongly advocated by Lindeberg [1991] and Weickert [1996] among others. The focus of the model analysis part is on necessary and sufficient conditions which guarantee that a discrete diffusion possesses the scale-space property in the sense of sign variation diminishing. Connections with the total variation diminishing and the open problem in a multivariate case are discussed too.
Considering the model selection, the thesis unifies two optimal diffusion stopping principles: (i) the time when the Shannon entropy-based Liapunov function of Sporring and Weickert [1999] reaches its steady state, and (ii) the time when the diffusion outcome has the least correlation with the noise estimate, contributed by Mrázek and Navara [2003]. Both ideas are shown to be particular cases of the marginal likelihood inference. Moreover, the suggested formalism provides first principles behind such criteria, and removes a variety of inconsistencies. It is suggested that the outcome of the diffusion should be interpreted as a certain expectation conditioned on the initial signal of observations instead of being treated as a random sample or probabilities. This removes the need to normalize signals in the approach of Sporring and Weickert [1999], and it also better justifies application of the correlation criterion of Mrázek and Navara [2003].
Throughout this work, the emphasis is given on methods that enable to reduce the problem to that of establishing the positivity of a quadratic form. The necessary and sufficient conditions can then be approached via positivity of matrix minors. A supplementary appendix is provided which summarizes a novel method of evaluating matrix minors. Intuitive examples of difficulties with statistical inference conclude the thesis.reviewe
Recent Advances in Wireless Communications and Networks
This book focuses on the current hottest issues from the lowest layers to the upper layers of wireless communication networks and provides "real-time" research progress on these issues. The authors have made every effort to systematically organize the information on these topics to make it easily accessible to readers of any level. This book also maintains the balance between current research results and their theoretical support. In this book, a variety of novel techniques in wireless communications and networks are investigated. The authors attempt to present these topics in detail. Insightful and reader-friendly descriptions are presented to nourish readers of any level, from practicing and knowledgeable communication engineers to beginning or professional researchers. All interested readers can easily find noteworthy materials in much greater detail than in previous publications and in the references cited in these chapters
- …