288 research outputs found
Hybrid system identification using switching density networks
Behaviour cloning is a commonly used strategy for imitation learning and can
be extremely effective in constrained domains. However, in cases where the
dynamics of an environment may be state dependent and varying, behaviour
cloning places a burden on model capacity and the number of demonstrations
required. This paper introduces switching density networks, which rely on a
categorical reparametrisation for hybrid system identification. This results in
a network comprising a classification layer that is followed by a regression
layer. We use switching density networks to predict the parameters of hybrid
control laws, which are toggled by a switching layer to produce different
controller outputs, when conditioned on an input state. This work shows how
switching density networks can be used for hybrid system identification in a
variety of tasks, successfully identifying the key joint angle goals that make
up manipulation tasks, while simultaneously learning image-based goal
classifiers and regression networks that predict joint angles from images. We
also show that they can cluster the phase space of an inverted pendulum,
identifying the balance, spin and pump controllers required to solve this task.
Switching density networks can be difficult to train, but we introduce a cross
entropy regularisation loss that stabilises training
DAC: The Double Actor-Critic Architecture for Learning Options
We reformulate the option framework as two parallel augmented MDPs. Under
this novel formulation, all policy optimization algorithms can be used off the
shelf to learn intra-option policies, option termination conditions, and a
master policy over options. We apply an actor-critic algorithm on each
augmented MDP, yielding the Double Actor-Critic (DAC) architecture.
Furthermore, we show that, when state-value functions are used as critics, one
critic can be expressed in terms of the other, and hence only one critic is
necessary. We conduct an empirical study on challenging robot simulation tasks.
In a transfer learning setting, DAC outperforms both its hierarchy-free
counterpart and previous gradient-based option learning algorithms.Comment: NeurIPS 201
Acquisition and distribution of synergistic reactive control skills
Learning from demonstration is an afficient way to attain a new skill. In the context of autonomous robots, using a demonstration to teach a robot accelerates the robot learning process significantly. It helps to identify feasible solutions as starting points for future exploration or to avoid actions that lead to failure. But the acquisition of pertinent observationa is predicated on first segmenting the data into meaningful sequences. These segments form the basis for learning models capable of recognising future actions and reconstructing the motion to control a robot. Furthermore, learning algorithms for generative models are generally not tuned to produce stable trajectories and suffer from parameter redundancy for high degree of freedom robots
This thesis addresses these issues by firstly investigating algorithms, based on dynamic programming and mixture models, for segmentation sensitivity and recognition accuracy on human motion capture data sets of repetitive and categorical motion classes. A stability analysis of the non-linear dynamical systems derived from the resultant mixture model representations aims to ensure that any trajectories converge to the intended target motion as observed in the demonstrations. Finally, these concepts are extended to humanoid robots by deploying a factor analyser for each mixture model component and coordinating the structure into a low dimensional representation of the demonstrated trajectories. This representation can be constructed as a correspondence map is learned between the demonstrator and robot for joint space actions.
Applying these algorithms for demonstrating movement skills to robot is a further step towards autonomous incremental robot learning
Bayesian Nonparametric Approaches for Modelling Stochastic Temporal Events
Modelling stochastic temporal events is a classic machine learning problem that has drawn enormous research attentions over recent decades. Traditional approaches heavily focused on the parametric models that pre-specify model complexity. Comprehensive model comparison and selection are necessary to prevent over-fitting and under-fitting problems.
The recently developed Bayesian nonparametric learning framework provides an appealing alternative to traditional approaches. It can automatically learn the model complexity from data. In this thesis, I propose a set of Bayesian nonparametric approaches for stochastic temporal event modelling with the consideration of event similarity, interaction, occurrence time and emitted observation. Specifically, I tackle following three main challenges in the modelling.
1. Data sparsity. Data sparsity problem is common in many real-world temporal event modelling applications, e.g., water pipes failures prediction. A Bayesian nonparametric model that allows pipes with similar behaviour to share failure data is proposed to attain a more effective failure prediction. It is shown that flexible event clustering can help alleviate the data sparsity problem. The clustering process is fully data-driven and it does not require predefining the number of clusters.
2. Event interaction. Stochastic events can interact with each other over time. One event can cause or repel the occurrence of other events. An unexplored theoretical bridge is established between interaction point processes and distance dependent Chinese restaurant process. Hence an integrated model, namely infinite branching model, is developed to estimate point event intensity, interaction mechanism and branching structure simultaneously.
3. Event correlation. The stochastic temporal events are correlated not only between arrival times but also between observations. A novel unified Bayesian nonparametric model that generalizes Hidden Markov model and interaction point processes is constructed to exploit two types of underlying correlation in a well-integrated way rather than individually. The proposed model provides a comprehensive insight into the interaction mechanism and correlation between events.
At last, a future vision of Bayesian nonparametric research for stochastic temporal events is highlighted from both application and modelling perspectives
Generative Models for Learning Robot Manipulation Skills from Humans
A long standing goal in artificial intelligence is to make robots seamlessly interact with humans in performing everyday manipulation skills. Learning from demonstrations or imitation learning provides a promising route to bridge this gap. In contrast to direct trajectory learning from demonstrations, many problems arise in interactive robotic applications that require higher contextual level understanding of the environment. This requires learning invariant mappings in the demonstrations that can generalize across different environmental situations such as size, position, orientation of objects, viewpoint of the observer, etc.
In this thesis, we address this challenge by encapsulating invariant patterns in the demonstrations using probabilistic learning models for acquiring dexterous manipulation skills. We learn the joint probability density function of the demonstrations with a hidden semi-Markov model, and smoothly follow the generated sequence of states with a linear quadratic tracking controller. The model exploits the invariant segments (also termed as sub-goals, options or actions) in the demonstrations and adapts the movement in accordance with the external environmental situations such as size, position and orientation of the objects in the environment using a task-parameterized formulation. We incorporate high-dimensional sensory data for skill acquisition by parsimoniously representing the demonstrations using statistical subspace clustering methods and exploit the coordination patterns in latent space. To adapt the models on the fly and/or teach new manipulation skills online with the streaming data, we formulate a non-parametric scalable online sequence clustering algorithm with Bayesian non-parametric mixture models to avoid the model selection problem while ensuring tractability under small variance asymptotics.
We exploit the developed generative models to perform manipulation skills with remotely operated vehicles over satellite communication in the presence of communication delays and limited bandwidth. A set of task-parameterized generative models are learned from the demonstrations of different manipulation skills provided by the teleoperator. The model captures the intention of teleoperator on one hand and provides assistance in performing remote manipulation tasks on the other hand under varying environmental situations. The assistance is formulated under time-independent shared control, where the model continuously corrects the remote arm movement based on the current state of the teleoperator; and/or time-dependent autonomous control, where the model synthesizes the movement of the remote arm for autonomous skill execution. Using the proposed methodology with the two-armed Baxter robot as a mock-up for semi-autonomous teleoperation, we are able to learn manipulation skills such as opening a valve, pick-and-place an object by obstacle avoidance, hot-stabbing (a specialized underwater task akin to peg-in-a-hole task), screw-driver target snapping, and tracking a carabiner in as few as 4 - 8 demonstrations. Our study shows that the proposed manipulation assistance formulations improve the performance of the teleoperator by reducing the task errors and the execution time, while catering for the environmental differences in performing remote manipulation tasks with limited bandwidth and communication delays
Towards Preemptive Text Edition using Topic Matching on Corpora
Nowadays, the results of scientific research are only recognized when published in papers for international
journals or magazines of the respective area of knowledge. This perspective reflects
the importance of having the work reviewed by peers. The revision encompasses a thorough
analysis on the work performed, including quality of writing and whether the study advances
the state-of-the-art, among other details. For these reasons, with the publishing of the document,
other researchers have an assurance of the high quality of the study presented and can,
therefore, make direct usage of the findings in their own work. The publishing of documents
creates a cycle of information exchange responsible for speeding up the progress behind the
development of new techniques, theories and technologies, resulting in added value for the
entire society.
Nonetheless, the existence of a detailed revision of the content sent for publication requires
additional effort and dedication from its authors. They must make sure that the manuscript is
of high quality, since sending a document with mistakes conveys an unprofessional image of the
authors, which may result in the rejection at the journal or magazine. The objective of this
work is to develop an algorithm capable of assisting in the writing of this type of documents, by
proposing suggestions of possible improvements or corrections according to its specific context.
The general idea for the solution proposed is for the algorithm to calculate suggestions of improvements
by comparing the content of the document being written in to that of similar published
documents on the field. In this context, a study on Natural Language Processing (NLP)
techniques used in the creation of models for representing the document and its subjects was
performed. NLP provides the tools for creating models to represent the documents and identify
their topics. The main concepts include n-grams and topic modeling. The study included also an
analysis of some works performed in the field of academic writing. The structure and contents
of this type of documents, the presentation of some of the characteristics that are common to
high quality articles, as well as the tools developed with the objective of helping in its writing
were also subject of analysis.
The developed algorithm derives from the combination of several tools backed up by a collection
of documents, as well as the logic connecting all components, implemented in the scope of this
Master’s. The collection of documents is constituted by full text of articles from different areas,
including Computer Science, Physics and Mathematics, among others. The topics of these documents
were extracted and stored in order to be fed to the algorithm. By comparing the topics
extracted from the document under analysis with those from the documents in the collection,
it is possible to select its closest documents, using them for the creation of suggestions. The algorithm is capable of proposing suggestions for word replacements which are more commonly
utilized in a given field of knowledge through a set of tools used in syntactic analysis, synonyms
search and morphological realization.
Both objective and subjective tests were conducted on the algorithm. They demonstrate that,
in some cases, the algorithm proposes suggestions which approximate the terms used in the document
to the most utilized terms in the state-of-the-art of a defined scientific field. This points
towards the idea that the usage of the algorithm should improve the quality of the documents,
as they become more similar to the ones already published. Even though the improvements to
the documents are minimal, they should be understood as a lower bound for the real utility of
the algorithm. This statement is partially justified by the existence of several parsing errors
both in the training and test sets, resulting from the parsing of the pdf files from the original
articles, which can be improved in a production system.
The main contributions of this work include the presentation of the study performed on the state
of the art, the design and implementation of the algorithm and the text editor developed as a
proof of concept. The analysis on the specificity of the context, which results from the tests
performed on different areas of knowledge, and the large collection of documents, gathered
during this Master’s program, are also important contributions of this work.Hoje em dia, a realização de uma investigação científica só é valorizada quando resulta na publicação
de artigos científicos em jornais ou revistas internacionais de renome na respetiva área
do conhecimento. Esta perspetiva reflete a importância de que os estudos realizados sejam
validados por pares. A validação implica uma análise detalhada do estudo realizado, incluindo
a qualidade da escrita e a existência de novidades, entre outros detalhes. Por estas razões,
com a publicação do documento, outros investigadores têm uma garantia de qualidade do estudo
realizado e podem, por isso, utilizar o conhecimento gerado para o seu próprio trabalho.
A publicação destes documentos cria um ciclo de troca de informação que é responsável por
acelerar o processo de desenvolvimento de novas técnicas, teorias e tecnologias, resultando na
produção de valor acrescido para a sociedade em geral.
Apesar de todas estas vantagens, a existência de uma verificação detalhada do conteúdo do
documento enviado para publicação requer esforço e trabalho acrescentado para os autores.
Estes devem assegurar-se da qualidade do manuscrito, visto que o envio de um documento
defeituoso transmite uma imagem pouco profissional dos autores, podendo mesmo resultar na
rejeição da sua publicação nessa revista ou ata de conferência. O objetivo deste trabalho é
desenvolver um algoritmo para ajudar os autores na escrita deste tipo de documentos, propondo
sugestões para melhoramentos tendo em conta o seu contexto específico.
A ideia genérica para solucionar o problema passa pela extração do tema do documento a ser
escrito, criando sugestões através da comparação do seu conteúdo com o de documentos científicos
antes publicados na mesma área. Tendo em conta esta ideia e o contexto previamente
apresentado, foi realizado um estudo de técnicas associadas à área de Processamento de Linguagem
Natural (PLN). O PLN fornece ferramentas para a criação de modelos capazes de representar
o documento e os temas que lhe estão associados. Os principais conceitos incluem n-grams e
modelação de tópicos (topic modeling). Para concluir o estudo, foram analisados trabalhos
realizados na área dos artigos científicos, estudando a sua estrutura e principais conteúdos,
sendo ainda abordadas algumas características comuns a artigos de qualidade e ferramentas
desenvolvidas para ajudar na sua escrita.
O algoritmo desenvolvido é formado pela junção de um conjunto de ferramentas e por uma
coleção de documentos, bem como pela lógica que liga todos os componentes, implementada
durante este trabalho de mestrado. Esta coleção de documentos é constituída por artigos completos
de algumas áreas, incluindo Informática, Física e Matemática, entre outras. Antes da
análise de documentos, foi feita a extração de tópicos da coleção utilizada. Deste forma, ao
extrair os tópicos do documento sob análise, é possível selecionar os documentos da coleção mais semelhantes, sendo estes utilizados para a criação de sugestões. Através de um conjunto de
ferramentas para análise sintática, pesquisa de sinónimos e realização morfológica, o algoritmo
é capaz de criar sugestões de substituições de palavras que são mais comummente utilizadas na
área.
Os testes realizados permitiram demonstrar que, em alguns casos, o algoritmo é capaz de fornecer
sugestões úteis de forma a aproximar os termos utilizados no documento com os termos
mais utilizados no estado de arte de uma determinada área científica. Isto constitui uma evidência
de que a utilização do algoritmo desenvolvido pode melhorar a qualidade da escrita de
documentos científicos, visto que estes tendem a aproximar-se daqueles já publicados. Apesar
dos resultados apresentados não refletirem uma grande melhoria no documento, estes deverão
ser considerados uma baixa estimativa ao valor real do algoritmo. Isto é justificado pela presença
de inúmeros erros resultantes da conversão dos documentos pdf para texto, estando estes
presentes tanto na coleção de documentos, como nos testes.
As principais contribuições deste trabalho incluem a partilha do estudo realizado, o desenho e
implementação do algoritmo e o editor de texto desenvolvido como prova de conceito. A análise
de especificidade de um contexto, que advém dos testes realizados às várias áreas do conhecimento,
e a extensa coleção de documentos, totalmente compilada durante este mestrado, são
também contribuições do trabalho
Recommended from our members
COMPUTATIONAL COMMUNICATION INTELLIGENCE: EXPLORING LINGUISTIC MANIFESTATION AND SOCIAL DYNAMICS IN ONLINE COMMUNICATION
We now live in an age of online communication. As social media becomes an integral part of our life, online communication becomes an essential life skill. In this dissertation, we aim to understand how people effectively communicate online. We research components of success in online communication and present scientific methods to study the skill of effective communication. This research advances the state of art in machine learning and communication studies.
For communication studies, we pioneer the study of a communication phenomenon we call Communication Intelligence in online interactions. We create a theory about communication intelligence that measures participants’ ten high-order communication skills, including restraint, self-reflection, perspective taking, and balance. We present a multi-perspective analysis for understanding communication intelligence, including its diverse language, shared linguistic characteristics across people, social dynamics, and the effects of communication modality on communication intelligence.
For machine learning, we contribute new computational models and formulations for addressing multi-label and multi-task machine learning problems. We develop a new hierarchical probabilistic model for simultaneously identifying multiple intelligence-embodied communication skills from natural language. The model learns the topic assignment for each sentence and provides a practical and simple way to determine document labels without relying on a threshold function. The model performance increases as the number of labels grows, which makes it a promising approach for large-scale data analysis. We also develop a new multi-task formulation for simultaneously identifying multiple intelligence-embodied communication skills from lexical, discourse, and interaction features. The key merit of this model is that it is a general multi-task formulation that unifies many widely used regularization techniques, including Lasso, group Lasso, sparse-group Lasso, and the Dirty model. This model expands the applicability of multi-task learning by allowing analyzing real-world problems where the degree of task relatedness is uncertain and the true structure of the groups in data is not clear ahead of time. Moreover, it can be applied to streaming data to perform large-scale analysis in real time. Beyond the application of studying communication intelligence, the developed models and formulations can also benefit research in other areas where the problems of simultaneously predicting multiple categories are abundant
- …