4,820 research outputs found
Lesson Learnt and Future of AI Applied to Manufacturing
This chapter touches on several aspects related to the role of Artificial Intelligence (AI) and Machine Learning (ML) in the manufacturing sector, and is split in different sub-chapters, focusing on specific new technology enablers that have the potential of solving or minimizing known issues in the manufacturing and, more in general, in the Industrial Internet of Things (IIoT) domain. After introducing AI/ML as a technology enabler for the IoT in general and for manufacturing in particular, the next four sections detail two key technology enablers (EdgeML and federated learning scenarios, challenges and tools), one most important area of the IoT system that needs to decrease energy consumption and increase reliability (reduce receiver Processing complexity and enhancing reliability through multi-connectivity uplink connections), and finally a glimpse at the future describing a promising new technology (Embodied AI), its link with millimetre waves connectivity and potential business impact
Towards edge robotics: the progress from cloud-based robotic systems to intelligent and context-aware robotic services
Current robotic systems handle a different range of applications such as video surveillance, delivery
of goods, cleaning, material handling, assembly, painting, or pick and place services. These systems
have been embraced not only by the general population but also by the vertical industries to
help them in performing daily activities. Traditionally, the robotic systems have been deployed in
standalone robots that were exclusively dedicated to performing a specific task such as cleaning the
floor in indoor environments. In recent years, cloud providers started to offer their infrastructures
to robotic systems for offloading some of the robot’s functions. This ultimate form of the distributed
robotic system was first introduced 10 years ago as cloud robotics and nowadays a lot of robotic solutions
are appearing in this form. As a result, standalone robots became software-enhanced objects
with increased reconfigurability as well as decreased complexity and cost. Moreover, by offloading
the heavy processing from the robot to the cloud, it is easier to share services and information from
various robots or agents to achieve better cooperation and coordination.
Cloud robotics is suitable for human-scale responsive and delay-tolerant robotic functionalities
(e.g., monitoring, predictive maintenance). However, there is a whole set of real-time robotic applications
(e.g., remote control, motion planning, autonomous navigation) that can not be executed with
cloud robotics solutions, mainly because cloud facilities traditionally reside far away from the robots.
While the cloud providers can ensure certain performance in their infrastructure, very little can be
ensured in the network between the robots and the cloud, especially in the last hop where wireless
radio access networks are involved. Over the last years advances in edge computing, fog computing,
5G NR, network slicing, Network Function Virtualization (NFV), and network orchestration are stimulating
the interest of the industrial sector to satisfy the stringent and real-time requirements of their
applications. Robotic systems are a key piece in the industrial digital transformation and their benefits
are very well studied in the literature. However, designing and implementing a robotic system
that integrates all the emerging technologies and meets the connectivity requirements (e.g., latency,
reliability) is an ambitious task.
This thesis studies the integration of modern Information andCommunication Technologies (ICTs)
in robotic systems and proposes some robotic enhancements that tackle the real-time constraints of
robotic services. To evaluate the performance of the proposed enhancements, this thesis departs
from the design and prototype implementation of an edge native robotic system that embodies the concepts of edge computing, fog computing, orchestration, and virtualization. The proposed edge
robotics system serves to represent two exemplary robotic applications. In particular, autonomous
navigation of mobile robots and remote-control of robot manipulator where the end-to-end robotic
system is distributed between the robots and the edge server. The open-source prototype implementation
of the designed edge native robotic system resulted in the creation of two real-world testbeds
that are used in this thesis as a baseline scenario for the evaluation of new innovative solutions in
robotic systems.
After detailing the design and prototype implementation of the end-to-end edge native robotic
system, this thesis proposes several enhancements that can be offered to robotic systems by adapting
the concept of edge computing via the Multi-Access Edge Computing (MEC) framework. First, it
proposes exemplary network context-aware enhancements in which the real-time information about
robot connectivity and location can be used to dynamically adapt the end-to-end system behavior to
the actual status of the communication (e.g., radio channel). Three different exemplary context-aware
enhancements are proposed that aim to optimize the end-to-end edge native robotic system. Later,
the thesis studies the capability of the edge native robotic system to offer potential savings by means of
computation offloading for robot manipulators in different deployment configurations. Further, the
impact of different wireless channels (e.g., 5G, 4G andWi-Fi) to support the data exchange between a
robot manipulator and its remote controller are assessed.
In the following part of the thesis, the focus is set on how orchestration solutions can support
mobile robot systems to make high quality decisions. The application of OKpi as an orchestration algorithm
and DLT-based federation are studied to meet the KPIs that autonomously controlledmobile
robots have in order to provide uninterrupted connectivity over the radio access network. The elaborated
solutions present high compatibility with the designed edge robotics system where the robot
driving range is extended without any interruption of the end-to-end edge robotics service. While the
DLT-based federation extends the robot driving range by deploying access point extension on top of
external domain infrastructure, OKpi selects the most suitable access point and computing resource
in the cloud-to-thing continuum in order to fulfill the latency requirements of autonomously controlled
mobile robots.
To conclude the thesis the focus is set on how robotic systems can improve their performance by
leveraging Artificial Intelligence (AI) and Machine Learning (ML) algorithms to generate smart decisions.
To do so, the edge native robotic system is presented as a true embodiment of a Cyber-Physical
System (CPS) in Industry 4.0, showing the mission of AI in such concept. It presents the key enabling
technologies of the edge robotic system such as edge, fog, and 5G, where the physical processes are
integrated with computing and network domains. The role of AI in each technology domain is identified
by analyzing a set of AI agents at the application and infrastructure level. In the last part of the
thesis, the movement prediction is selected to study the feasibility of applying a forecast-based recovery
mechanism for real-time remote control of robotic manipulators (FoReCo) that uses ML to infer
lost commands caused by interference in the wireless channel. The obtained results are showcasing
the its potential in simulation and real-world experimentation.Programa de Doctorado en IngenierĂa Telemática por la Universidad Carlos III de MadridPresidente: Karl Holger.- Secretario: Joerg Widmer.- Vocal: Claudio Cicconett
Co-design of Security Aware Power System Distribution Architecture as Cyber Physical System
The modern smart grid would involve deep integration between measurement nodes, communication systems, artificial intelligence, power electronics and distributed resources. On one hand, this type of integration can dramatically improve the grid performance and efficiency, but on the other, it can also introduce new types of vulnerabilities to the grid. To obtain the best performance, while minimizing the risk of vulnerabilities, the physical power system must be designed as a security aware system.
In this dissertation, an interoperability and communication framework for microgrid control and Cyber Physical system enhancements is designed and implemented taking into account cyber and physical security aspects. The proposed data-centric interoperability layer provides a common data bus and a resilient control network for seamless integration of distributed energy resources. In addition, a synchronized measurement network and advanced metering infrastructure were developed to provide real-time monitoring for active distribution networks.
A hybrid hardware/software testbed environment was developed to represent the smart grid as a cyber-physical system through hardware and software in the loop simulation methods. In addition it provides a flexible interface for remote integration and experimentation of attack scenarios.
The work in this dissertation utilizes communication technologies to enhance the performance of the DC microgrids and distribution networks by extending the application of the GPS synchronization to the DC Networks. GPS synchronization allows the operation of distributed DC-DC converters as an interleaved converters system. Along with the GPS synchronization, carrier extraction synchronization technique was developed to improve the system’s security and reliability in the case of GPS signal spoofing or jamming.
To improve the integration of the microgrid with the utility system, new synchronization and islanding detection algorithms were developed. The developed algorithms overcome the problem of SCADA and PMU based islanding detection methods such as communication failure and frequency stability. In addition, a real-time energy management system with online optimization was developed to manage the energy resources within the microgrid. The security and privacy were also addressed in both the cyber and physical levels. For the physical design, two techniques were developed to address the physical privacy issues by changing the current and electromagnetic signature. For the cyber level, a security mechanism for IEC 61850 GOOSE messages was developed to address the security shortcomings in the standard
High-Performance Modelling and Simulation for Big Data Applications
This open access book was prepared as a Final Publication of the COST Action IC1406 “High-Performance Modelling and Simulation for Big Data Applications (cHiPSet)“ project. Long considered important pillars of the scientific method, Modelling and Simulation have evolved from traditional discrete numerical methods to complex data-intensive continuous analytical optimisations. Resolution, scale, and accuracy have become essential to predict and analyse natural and complex systems in science and engineering. When their level of abstraction raises to have a better discernment of the domain at hand, their representation gets increasingly demanding for computational and data resources. On the other hand, High Performance Computing typically entails the effective use of parallel and distributed processing units coupled with efficient storage, communication and visualisation systems to underpin complex data-intensive applications in distinct scientific and technical domains. It is then arguably required to have a seamless interaction of High Performance Computing with Modelling and Simulation in order to store, compute, analyse, and visualise large data sets in science and engineering. Funded by the European Commission, cHiPSet has provided a dynamic trans-European forum for their members and distinguished guests to openly discuss novel perspectives and topics of interests for these two communities. This cHiPSet compendium presents a set of selected case studies related to healthcare, biological data, computational advertising, multimedia, finance, bioinformatics, and telecommunications
Structure, Dynamics and Self-Organization in Recurrent Neural Networks: From Machine Learning to Theoretical Neuroscience
At a first glance, artificial neural networks, with engineered learning algorithms and carefully chosen nonlinearities, are nothing like the complicated self-organized spiking neural networks studied by theoretical neuroscientists. Yet, both adapt to their inputs, keep information from the past in their state space and are able of learning, implying that some information processing principles should be common to both. In this thesis we study those principles by incorporating notions of systems theory, statistical physics and graph theory into artificial neural networks and theoretical neuroscience models.
% TO DO: What is different in this thesis? -> classical signal processing with complex systems on top
The starting point for this thesis is \ac{RC}, a learning paradigm used both in machine learning\cite{jaeger2004harnessing} and in theoretical neuroscience\cite{maass2002real}. A neural network in \ac{RC} consists of two parts, a reservoir – a directed and weighted network of neurons that projects the input time series onto a high dimensional space – and a readout which is trained to read the state of the neurons in the reservoir and combine them linearly to give the desired output. In classical \ac{RC}, the reservoir is randomly initialized and left untrained, which alleviates the training costs in comparison to other recurrent neural networks. However, this lack of training implies that reservoirs are not adapted to specific tasks and thus their performance is often lower than that of other neural networks. Our contribution has been to show how knowledge about a task can be integrated into the reservoir architecture, so that reservoirs can be tailored to specific problems without training.
We do this design by identifying two features that are useful for machine learning: the memory of the reservoir and its power spectra. First we show that the correlations between neurons limit the capacity of the reservoir to retain traces of previous inputs, and demonstrate that those correlations are controlled by moduli of the eigenvalues of the adjacency matrix of the reservoir. Second, we prove that when the reservoir resonates at the frequencies that are present on the desired output signal, the performance of the readout increases.
Knowing the features of the reservoir dynamics that we need, the next question is how to impose them. The simplest way to design a network with that resonates at a certain frequency is by adding cycles, which act as feedback loops, but this also induces correlations and hence memory modifications. To disentangle the frequencies and the memory design, we studied how the addition of cycles modifies the eigenvalues in the adjacency matrix of the network. Surprisingly, the shape of the eigenvalues is quite beautiful \cite{aceituno2019universal} and can be characterized using random matrix theory tools. Combining this knowledge with our result relating eigenvalues and correlations, we designed an heuristic that tailors reservoirs to specific tasks and showed that it improves upon state of the art \ac{RC} in three different machine learning tasks.
Although this idea works in the machine learning version of \ac{RC}, there is one fundamental problem when we try to translate to the world of theoretical neuroscience: the proposed frequency adaptation requires prior knowledge of the task, which might not be plausible in a biological neural network. Therefore the following questions are whether those resonances can emerge by unsupervised learning, and which kind of learning rules would be required.
Remarkably, these resonances can be induced by the well-known Spike Time-Dependent Plasticity (STDP) combined with homeostatic mechanisms. We show this by deriving two self-consistent equations: one where the activity of every neuron can be calculated from its synaptic weights and its external inputs and a second one where the synaptic weights can be obtained from the neural activity. By considering spatio-temporal symmetries in our inputs we obtained two families of solutions to those equations where a periodic input is enhanced by the neural network after STDP. This approach shows that periodic and quasiperiodic inputs can induce resonances that agree with the aforementioned \ac{RC} theory.
Those results, although rigorous, are expressed on a language of statistical physics and cannot be easily tested or verified in real, scarce data. To make them more accessible to the neuroscience community we showed that latency reduction, a well-known effect of STDP\cite{song2000competitive} which has been experimentally observed \cite{mehta2000experience}, generates neural codes that agree with the self-consistency equations and their solutions. In particular, this analysis shows that metabolic efficiency, synchronization and predictions can emerge from that same phenomena of latency reduction, thus closing the loop with our original machine learning problem.
To summarize, this thesis exposes principles of learning recurrent neural networks that are consistent with adaptation in the nervous system and also improve current machine learning methods. This is done by leveraging features of the dynamics of recurrent neural networks such as resonances and correlations in machine learning problems, then imposing the required dynamics into reservoir computing through control theory notions such as feedback loops and spectral analysis. Then we assessed the plausibility of such adaptation in biological networks, deriving solutions from self-organizing processes that are biologically plausible and align with the machine learning prescriptions. Finally, we relate those processes to learning rules in biological neurons, showing how small local adaptations of the spike times can lead to neural codes that are efficient and can be interpreted in machine learning terms
Big Data and the Internet of Things
Advances in sensing and computing capabilities are making it possible to
embed increasing computing power in small devices. This has enabled the sensing
devices not just to passively capture data at very high resolution but also to
take sophisticated actions in response. Combined with advances in
communication, this is resulting in an ecosystem of highly interconnected
devices referred to as the Internet of Things - IoT. In conjunction, the
advances in machine learning have allowed building models on this ever
increasing amounts of data. Consequently, devices all the way from heavy assets
such as aircraft engines to wearables such as health monitors can all now not
only generate massive amounts of data but can draw back on aggregate analytics
to "improve" their performance over time. Big data analytics has been
identified as a key enabler for the IoT. In this chapter, we discuss various
avenues of the IoT where big data analytics either is already making a
significant impact or is on the cusp of doing so. We also discuss social
implications and areas of concern.Comment: 33 pages. draft of upcoming book chapter in Japkowicz and Stefanowski
(eds.) Big Data Analysis: New algorithms for a new society, Springer Series
on Studies in Big Data, to appea
Past, Present, and Future of Simultaneous Localization And Mapping: Towards the Robust-Perception Age
Simultaneous Localization and Mapping (SLAM)consists in the concurrent
construction of a model of the environment (the map), and the estimation of the
state of the robot moving within it. The SLAM community has made astonishing
progress over the last 30 years, enabling large-scale real-world applications,
and witnessing a steady transition of this technology to industry. We survey
the current state of SLAM. We start by presenting what is now the de-facto
standard formulation for SLAM. We then review related work, covering a broad
set of topics including robustness and scalability in long-term mapping, metric
and semantic representations for mapping, theoretical performance guarantees,
active SLAM and exploration, and other new frontiers. This paper simultaneously
serves as a position paper and tutorial to those who are users of SLAM. By
looking at the published research with a critical eye, we delineate open
challenges and new research issues, that still deserve careful scientific
investigation. The paper also contains the authors' take on two questions that
often animate discussions during robotics conferences: Do robots need SLAM? and
Is SLAM solved
- …