47 research outputs found
Kimera: from SLAM to Spatial Perception with 3D Dynamic Scene Graphs
Humans are able to form a complex mental model of the environment they move
in. This mental model captures geometric and semantic aspects of the scene,
describes the environment at multiple levels of abstractions (e.g., objects,
rooms, buildings), includes static and dynamic entities and their relations
(e.g., a person is in a room at a given time). In contrast, current robots'
internal representations still provide a partial and fragmented understanding
of the environment, either in the form of a sparse or dense set of geometric
primitives (e.g., points, lines, planes, voxels) or as a collection of objects.
This paper attempts to reduce the gap between robot and human perception by
introducing a novel representation, a 3D Dynamic Scene Graph(DSG), that
seamlessly captures metric and semantic aspects of a dynamic environment. A DSG
is a layered graph where nodes represent spatial concepts at different levels
of abstraction, and edges represent spatio-temporal relations among nodes. Our
second contribution is Kimera, the first fully automatic method to build a DSG
from visual-inertial data. Kimera includes state-of-the-art techniques for
visual-inertial SLAM, metric-semantic 3D reconstruction, object localization,
human pose and shape estimation, and scene parsing. Our third contribution is a
comprehensive evaluation of Kimera in real-life datasets and photo-realistic
simulations, including a newly released dataset, uHumans2, which simulates a
collection of crowded indoor and outdoor scenes. Our evaluation shows that
Kimera achieves state-of-the-art performance in visual-inertial SLAM, estimates
an accurate 3D metric-semantic mesh model in real-time, and builds a DSG of a
complex indoor environment with tens of objects and humans in minutes. Our
final contribution shows how to use a DSG for real-time hierarchical semantic
path-planning. The core modules in Kimera are open-source.Comment: 34 pages, 25 figures, 9 tables. arXiv admin note: text overlap with
arXiv:2002.0628
Emotion and Stress Recognition Related Sensors and Machine Learning Technologies
This book includes impactful chapters which present scientific concepts, frameworks, architectures and ideas on sensing technologies and machine learning techniques. These are relevant in tackling the following challenges: (i) the field readiness and use of intrusive sensor systems and devices for capturing biosignals, including EEG sensor systems, ECG sensor systems and electrodermal activity sensor systems; (ii) the quality assessment and management of sensor data; (iii) data preprocessing, noise filtering and calibration concepts for biosignals; (iv) the field readiness and use of nonintrusive sensor technologies, including visual sensors, acoustic sensors, vibration sensors and piezoelectric sensors; (v) emotion recognition using mobile phones and smartwatches; (vi) body area sensor networks for emotion and stress studies; (vii) the use of experimental datasets in emotion recognition, including dataset generation principles and concepts, quality insurance and emotion elicitation material and concepts; (viii) machine learning techniques for robust emotion recognition, including graphical models, neural network methods, deep learning methods, statistical learning and multivariate empirical mode decomposition; (ix) subject-independent emotion and stress recognition concepts and systems, including facial expression-based systems, speech-based systems, EEG-based systems, ECG-based systems, electrodermal activity-based systems, multimodal recognition systems and sensor fusion concepts and (x) emotion and stress estimation and forecasting from a nonlinear dynamical system perspective
Recommended from our members
ReSCon '12, Research Student Conference: Book of Abstracts
The fifth SED Research Student Conference (ReSCon2012) was hosted over three days, 18-20 June 2012, in the Hamilton Centre at Brunel University. The conference consisted of 130 oral and 70 poster presentations, based on the high quality and diverse research being conducted within the School of Engineering and Design by postgraduate research students. The conference is held annually, and ReSCon plays a key role in contributing to research and innovations within the School
Discriminative, generative, and imitative learning
Thesis (Ph. D.)--Massachusetts Institute of Technology, School of Architecture and Planning, Program in Media Arts and Sciences, 2002.Includes bibliographical references (leaves 201-212).I propose a common framework that combines three different paradigms in machine learning: generative, discriminative and imitative learning. A generative probabilistic distribution is a principled way to model many machine learning and machine perception problems. Therein, one provides domain specific knowledge in terms of structure and parameter priors over the joint space of variables. Bayesian networks and Bayesian statistics provide a rich and flexible language for specifying this knowledge and subsequently refining it with data and observations. The final result is a distribution that is a good generator of novel exemplars. Conversely, discriminative algorithms adjust a possibly non-distributional model to data optimizing for a specific task, such as classification or prediction. This typically leads to superior performance yet compromises the flexibility of generative modeling. I present Maximum Entropy Discrimination (MED) as a framework to combine both discriminative estimation and generative probability densities. Calculations involve distributions over parameters, margins, and priors and are provably and uniquely solvable for the exponential family. Extensions include regression, feature selection, and transduction. SVMs are also naturally subsumed and can be augmented with, for example, feature selection, to obtain substantial improvements. To extend to mixtures of exponential families, I derive a discriminative variant of the Expectation-Maximization (EM) algorithm for latent discriminative learning (or latent MED).(cont.) While EM and Jensen lower bound log-likelihood, a dual upper bound is made possible via a novel reverse-Jensen inequality. The variational upper bound on latent log-likelihood has the same form as EM bounds, is computable efficiently and is globally guaranteed. It permits powerful discriminative learning with the wide range of contemporary probabilistic mixture models (mixtures of Gaussians, mixtures of multinomials and hidden Markov models). We provide empirical results on standardized data sets that demonstrate the viability of the hybrid discriminative-generative approaches of MED and reverse-Jensen bounds over state of the art discriminative techniques or generative approaches. Subsequently, imitative learning is presented as another variation on generative modeling which also learns from exemplars from an observed data source. However, the distinction is that the generative model is an agent that is interacting in a much more complex surrounding external world. It is not efficient to model the aggregate space in a generative setting. I demonstrate that imitative learning (under appropriate conditions) can be adequately addressed as a discriminative prediction task which outperforms the usual generative approach. This discriminative-imitative learning approach is applied with a generative perceptual system to synthesize a real-time agent that learns to engage in social interactive behavior.by Tony Jebara.Ph.D
Validação de heterogeneidade estrutural em dados de Crio-ME por comitês de agrupadores
Orientadores: Fernando JosĂ© Von Zuben, Rodrigo Villares PortugalDissertação (mestrado) - Universidade Estadual de Campinas, Faculdade de Engenharia ElĂ©trica e de ComputaçãoResumo: Análise de PartĂculas Isoladas Ă© uma tĂ©cnica que permite o estudo da estrutura tridimensional de proteĂnas e outros complexos macromoleculares de interesse biolĂłgico. Seus dados primários consistem em imagens de microscopia eletrĂ´nica de transmissĂŁo de mĂşltiplas cĂłpias da molĂ©cula em orientações aleatĂłrias. Tais imagens sĂŁo bastante ruidosas devido Ă baixa dose de elĂ©trons utilizada. Reconstruções 3D podem ser obtidas combinando-se muitas imagens de partĂculas em orientações similares e estimando seus ângulos relativos. Entretanto, estados conformacionais heterogĂŞneos frequentemente coexistem na amostra, porque os complexos moleculares podem ser flexĂveis e tambĂ©m interagir com outras partĂculas. Heterogeneidade representa um desafio na reconstrução de modelos 3D confiáveis e degrada a resolução dos mesmos. Entre os algoritmos mais populares usados para classificação estrutural estĂŁo o agrupamento por k-mĂ©dias, agrupamento hierárquico, mapas autoorganizáveis e estimadores de máxima verossimilhança. Tais abordagens estĂŁo geralmente entrelaçadas Ă reconstrução dos modelos 3D. No entanto, trabalhos recentes indicam ser possĂvel inferir informações a respeito da estrutura das molĂ©culas diretamente do conjunto de projeções 2D. Dentre estas descobertas, está a relação entre a variabilidade estrutural e manifolds em um espaço de atributos multidimensional. Esta dissertação investiga se um comitĂŞ de algoritmos de nĂŁo-supervisionados Ă© capaz de separar tais "manifolds conformacionais". MĂ©todos de "consenso" tendem a fornecer classificação mais precisa e podem alcançar performance satisfatĂłria em uma ampla gama de conjuntos de dados, se comparados a algoritmos individuais. NĂłs investigamos o comportamento de seis algoritmos de agrupamento, tanto individualmente quanto combinados em comitĂŞs, para a tarefa de classificação de heterogeneidade conformacional. A abordagem proposta foi testada em conjuntos sintĂ©ticos e reais contendo misturas de imagens de projeção da proteĂna Mm-cpn nos estados "aberto" e "fechado". Demonstra-se que comitĂŞs de agrupadores podem fornecer informações Ăşteis na validação de particionamentos estruturais independetemente de algoritmos de reconstrução 3DAbstract: Single Particle Analysis is a technique that allows the study of the three-dimensional structure of proteins and other macromolecular assemblies of biological interest. Its primary data consists of transmission electron microscopy images from multiple copies of the molecule in random orientations. Such images are very noisy due to the low electron dose employed. Reconstruction of the macromolecule can be obtained by averaging many images of particles in similar orientations and estimating their relative angles. However, heterogeneous conformational states often co-exist in the sample, because the molecular complexes can be flexible and may also interact with other particles. Heterogeneity poses a challenge to the reconstruction of reliable 3D models and degrades their resolution. Among the most popular algorithms used for structural classification are k-means clustering, hierarchical clustering, self-organizing maps and maximum-likelihood estimators. Such approaches are usually interlaced with the reconstructions of the 3D models. Nevertheless, recent works indicate that it is possible to infer information about the structure of the molecules directly from the dataset of 2D projections. Among these findings is the relationship between structural variability and manifolds in a multidimensional feature space. This dissertation investigates whether an ensemble of unsupervised classification algorithms is able to separate these "conformational manifolds". Ensemble or "consensus" methods tend to provide more accurate classification and may achieve satisfactory performance across a wide range of datasets, when compared with individual algorithms. We investigate the behavior of six clustering algorithms both individually and combined in ensembles for the task of structural heterogeneity classification. The approach was tested on synthetic and real datasets containing a mixture of images from the Mm-cpn chaperonin in the "open" and "closed" states. It is shown that cluster ensembles can provide useful information in validating the structural partitionings independently of 3D reconstruction methodsMestradoEngenharia de ComputaçãoMestre em Engenharia ElĂ©tric
Building the Future Internet through FIRE
The Internet as we know it today is the result of a continuous activity for improving network communications, end user services, computational processes and also information technology infrastructures. The Internet has become a critical infrastructure for the human-being by offering complex networking services and end-user applications that all together have transformed all aspects, mainly economical, of our lives. Recently, with the advent of new paradigms and the progress in wireless technology, sensor networks and information systems and also the inexorable shift towards everything connected paradigm, first as known as the Internet of Things and lately envisioning into the Internet of Everything, a data-driven society has been created. In a data-driven society, productivity, knowledge, and experience are dependent on increasingly open, dynamic, interdependent and complex Internet services. The challenge for the Internet of the Future design is to build robust enabling technologies, implement and deploy adaptive systems, to create business opportunities considering increasing uncertainties and emergent systemic behaviors where humans and machines seamlessly cooperate
From Network to Web dimension in supply chain management
Cette thèse soutient que la dimension réseau, étant actuellement la portée du domaine de la gestion de chaîne logistique, contraint l’avancement de ce domaine et restreint des innovations conceptuelles et fondamentales capables d’adresser les grands défis économiques, environnementaux et sociaux. Les concepts de chaîne et de réseau ne reflètent pas la complexité des flux physiques, informationnels et financiers générés par les interactions qui ont lieu dans des réseaux interconnectés. Ces concepts n’offrent pas les fondations théoriques pour supporter des interventions allant au-delà d’un seul réseau et laissent échapper des opportunités nécessitant une vision multi-réseau. Ainsi, la dimension “web”, celle des réseaux de réseaux, est proposée comme une extension de la dimension réseau. Cette extension peut être vue comme l’étape naturelle suivante dans la progression qui a commencé par le niveau de gestion des opérations internes, est passée au niveau de la chaîne logistique et se trouve actuellement au niveau du réseau logistique. Après l’investigation théorique des raisons et de la façon d’intégrer la dimension web dans le domaine de la gestion de la chaîne logistique, la thèse étudie des implications importantes de cette intégration sur la collaboration inter-organisationnelle et le processus de prise de décision dans des environnements de webs logistiques. Elle démontre, en exploitant l’exemple des réseaux interconnectés ouverts, des potentialités inimaginables sans une vision web. Une méthodologie de conception d’un modèle de simulation permettant l’évaluation et la comparaison des webs ouverts par rapport aux webs existants est proposée. Puisque l’aide à la décision est une composante importante de la gestion de la chaîne logistique, la thèse contribue à déterminer les besoins des gestionnaires et à identifier les lignes directrices de la conception des outils d’aide à la décision offrant le support adéquat pour faire face aux défis et à la complexité des webs logistiques. Ces lignes directrices ont été compilées dans un cadre de conception des logiciels d’aide à la décision supportant la dimension web. Ce cadre est exploité pour développer quatre applications logicielles offrant aux praticiens et aux chercheurs des outils nécessaires pour étudier, analyser et démêler la complexité des webs logistiques.This thesis argues that the network dimension as the current scope of supply chain management is confining the evolution of this field and restricting the conceptual and fundamental innovations required for addressing the major challenges imposed by the evolution of markets and the increased intricacies of business relationships. The concepts of chain and network are limitative when attempting to represent the complexity of physical, informational and financial flows resulting from the interactions occurring in overlapping networks. They lack the theoretical foundations necessary to explain and encompass initiatives that go beyond a single chain or network. They also lead to overlook substantial opportunities that require beyond a network vision. Therefore, the “web” dimension, as networks of networks, is proposed as an extension to the network dimension in supply chain management. This new scope is the natural next step in the progression from the internal operations management level to the supply chain level and then to the supply network level. After a theoretical investigation of why and how the web dimension should be integrated into the supply chain management field, the thesis studies and discusses important implications of this integration on inter-organisational collaboration and of the decision-making processes in the logistic web environments. It demonstrates through the example of open interconnected logistic webs some of the potentials that cannot be imagined without a web vision. A methodology for designing a simulation model to assess the impact of such open webs versus existing webs is proposed. Since decision support is a key element in supply chain management, the thesis contributes to determine the needs of supply chain managers and identify the important axes for designing decision support systems that provide adequate assistance in dealing with the challenges and complexity presented by logistic web environments. The identified elements result in the establishment of a foundation for designing software solutions required to handle the challenges revealed by the web dimension. This conceptual framework is applied to the prototyping of four applications that have the potential of providing practitioners and researchers with the appropriate understanding and necessary tools to deal with the complexity of logistics webs