413 research outputs found
Detecting and Classifying Human Touches in a Social Robot Through Acoustic Sensing and Machine Learning
An important aspect in Human-Robot Interaction is responding to different kinds of touch stimuli. To date, several technologies have been explored to determine how a touch is perceived by a social robot, usually placing a large number of sensors throughout the robot's shell. In this work, we introduce a novel approach, where the audio acquired from contact microphones located in the robot's shell is processed using machine learning techniques to distinguish between different types of touches. The system is able to determine when the robot is touched (touch detection), and to ascertain the kind of touch performed among a set of possibilities: stroke, tap, slap, and tickle (touch classification). This proposal is cost-effective since just a few microphones are able to cover the whole robot's shell since a single microphone is enough to cover each solid part of the robot. Besides, it is easy to install and configure as it just requires a contact surface to attach the microphone to the robot's shell and plug it into the robot's computer. Results show the high accuracy scores in touch gesture recognition. The testing phase revealed that Logistic Model Trees achieved the best performance, with an F-score of 0.81. The dataset was built with information from 25 participants performing a total of 1981 touch gestures.The research leading to these results has received funding from the projects: Development of social robots to help seniors with cognitive impairment (ROBSEN), funded by the Ministerio de Economia y Competitividad; and RoboCity2030-III-CM, funded by Comunidad de Madrid and cofunded by Structural Funds of the EU.Publicad
Detecting, locating and recognising human touches in social robots with contact microphones
There are many situations in our daily life where touch gestures during natural human–human interaction take place: meeting people (shaking hands), personal relationships (caresses), moments of celebration or sadness (hugs), etc. Considering that robots are expected to form part of our daily life in the future, they should be endowed with the capacity of recognising these touch gestures and the part of its body that has been touched since the gesture’s meaning may differ. Therefore, this work presents a learning system for both purposes: detect and recognise the type of touch gesture (stroke, tickle, tap and slap) and its localisation. The interpretation of the meaning of the gesture is out of the scope of this paper.
Different technologies have been applied to perceive touch by a social robot, commonly using a large number of sensors. Instead, our approach uses 3 contact microphones installed inside some parts of the robot. The audio signals generated when the user touches the robot are sensed by the contact microphones and processed using Machine Learning techniques. We acquired information from sensors installed in two social robots, Maggie and Mini (both developed by the RoboticsLab at the Carlos III University of Madrid), and a real-time version of the whole system has been deployed in the robot Mini. The system allows the robot to sense if it has been touched or not, to recognise the kind of touch gesture, and its approximate location. The main advantage of using contact microphones as touch sensors is that by using just one, it is possible to “cover” a whole solid part of the robot. Besides, the sensors are unaffected by ambient noises, such as human voice, TV, music etc. Nevertheless, the fact of using several contact microphones makes possible that a touch gesture is detected by all of them, and each may recognise a different gesture at the same time. The results show that this system is robust against this phenomenon. Moreover, the accuracy obtained for both robots is about 86%.The research leading to these results has received funding from
the projects: ‘‘Robots Sociales para Estimulación Física, Cognitiva y
Afectiva de Mayores (ROSES)’’, funded by the Spanish "Ministerio de
Ciencia, Innovación y Universidades, Spain" and from RoboCity2030-DIH-CM, Madrid Robotics Digital Innovation Hub, S2018/NMT-4331,
funded by ‘"Programas de Actividades I+D en la Comunidad de Madrid’"
and cofunded by Structural Funds of the EU, Slovak Republic.Publicad
Acoustic-based Smart Tactile Sensing in Social Robots
Mención Internacional en el título de doctorEl sentido del tacto es un componente crucial de la interacción social humana y es único
entre los cinco sentidos. Como único sentido proximal, el tacto requiere un contacto
físico cercano o directo para registrar la información. Este hecho convierte al tacto en
una modalidad de interacción llena de posibilidades en cuanto a comunicación social. A través
del tacto, podemos conocer la intención de la otra persona y comunicar emociones. De esta
idea surge el concepto de social touch o tacto social como el acto de tocar a otra persona en
un contexto social. Puede servir para diversos fines, como saludar, mostrar afecto, persuadir
y regular el bienestar emocional y físico.
Recientemente, el número de personas que interactúan con sistemas y agentes artificiales
ha aumentado, principalmente debido al auge de los dispositivos tecnológicos, como los smartphones
o los altavoces inteligentes. A pesar del auge de estos dispositivos, sus capacidades de
interacción son limitadas. Para paliar este problema, los recientes avances en robótica social han
mejorado las posibilidades de interacción para que los agentes funcionen de forma más fluida y
sean más útiles. En este sentido, los robots sociales están diseñados para facilitar interacciones
naturales entre humanos y agentes artificiales. El sentido del tacto en este contexto se revela
como un vehículo natural que puede mejorar la Human-Robot Interaction (HRI) debido a su
relevancia comunicativa en entornos sociales. Además de esto, para un robot social, la relación
entre el tacto social y su aspecto es directa, al disponer de un cuerpo físico para aplicar o recibir
toques.
Desde un punto de vista técnico, los sistemas de detección táctil han sido objeto recientemente
de nuevas investigaciones, sobre todo dedicado a comprender este sentido para crear sistemas
inteligentes que puedan mejorar la vida de las personas. En este punto, los robots sociales
se han convertido en dispositivos muy populares que incluyen tecnologías para la detección
táctil. Esto está motivado por el hecho de que un robot puede esperada o inesperadamente
tener contacto físico con una persona, lo que puede mejorar o interferir en la ejecución de sus
comportamientos. Por tanto, el sentido del tacto se antoja necesario para el desarrollo de aplicaciones
robóticas. Algunos métodos incluyen el reconocimiento de gestos táctiles, aunque
a menudo exigen importantes despliegues de hardware que requieren de múltiples sensores. Además, la fiabilidad de estas tecnologías de detección es limitada, ya que la mayoría de ellas
siguen teniendo problemas tales como falsos positivos o tasas de reconocimiento bajas. La detección
acústica, en este sentido, puede proporcionar un conjunto de características capaces de
paliar las deficiencias anteriores. A pesar de que se trata de una tecnología utilizada en diversos
campos de investigación, aún no se ha integrado en la interacción táctil entre humanos y robots.
Por ello, en este trabajo proponemos el sistema Acoustic Touch Recognition (ATR), un sistema
inteligente de detección táctil (smart tactile sensing system) basado en la detección acústica
y diseñado para mejorar la interacción social humano-robot. Nuestro sistema está desarrollado
para clasificar gestos táctiles y localizar su origen. Además de esto, se ha integrado en plataformas
robóticas sociales y se ha probado en aplicaciones reales con éxito. Nuestra propuesta
se ha enfocado desde dos puntos de vista: uno técnico y otro relacionado con el tacto social.
Por un lado, la propuesta tiene una motivación técnica centrada en conseguir un sistema táctil
rentable, modular y portátil. Para ello, en este trabajo se ha explorado el campo de las tecnologías
de detección táctil, los sistemas inteligentes de detección táctil y su aplicación en HRI. Por
otro lado, parte de la investigación se centra en el impacto afectivo del tacto social durante la
interacción humano-robot, lo que ha dado lugar a dos estudios que exploran esta idea.The sense of touch is a crucial component of human social interaction and is unique
among the five senses. As the only proximal sense, touch requires close or direct physical
contact to register information. This fact makes touch an interaction modality
full of possibilities regarding social communication. Through touch, we are able to ascertain
the other person’s intention and communicate emotions. From this idea emerges the concept
of social touch as the act of touching another person in a social context. It can serve various purposes,
such as greeting, showing affection, persuasion, and regulating emotional and physical
well-being.
Recently, the number of people interacting with artificial systems and agents has increased,
mainly due to the rise of technological devices, such as smartphones or smart speakers. Still,
these devices are limited in their interaction capabilities. To deal with this issue, recent developments
in social robotics have improved the interaction possibilities to make agents more seamless
and useful. In this sense, social robots are designed to facilitate natural interactions between
humans and artificial agents. In this context, the sense of touch is revealed as a natural interaction
vehicle that can improve HRI due to its communicative relevance. Moreover, for a social
robot, the relationship between social touch and its embodiment is direct, having a physical
body to apply or receive touches.
From a technical standpoint, tactile sensing systems have recently been the subject of further
research, mostly devoted to comprehending this sense to create intelligent systems that can
improve people’s lives. Currently, social robots are popular devices that include technologies
for touch sensing. This is motivated by the fact that robots may encounter expected or unexpected
physical contact with humans, which can either enhance or interfere with the execution
of their behaviours. There is, therefore, a need to detect human touch in robot applications.
Some methods even include touch-gesture recognition, although they often require significant
hardware deployments primarily that require multiple sensors. Additionally, the dependability
of those sensing technologies is constrained because the majority of them still struggle with issues
like false positives or poor recognition rates. Acoustic sensing, in this sense, can provide a
set of features that can alleviate the aforementioned shortcomings. Even though it is a technology that has been utilised in various research fields, it has yet to be integrated into human-robot
touch interaction.
Therefore, in thiswork,we propose theATRsystem, a smart tactile sensing system based on
acoustic sensing designed to improve human-robot social interaction. Our system is developed
to classify touch gestures and locate their source. It is also integrated into real social robotic platforms
and tested in real-world applications. Our proposal is approached from two standpoints,
one technical and the other related to social touch. Firstly, the technical motivation of thiswork
centred on achieving a cost-efficient, modular and portable tactile system. For that, we explore
the fields of touch sensing technologies, smart tactile sensing systems and their application in
HRI. On the other hand, part of the research is centred around the affective impact of touch
during human-robot interaction, resulting in two studies exploring this idea.Programa de Doctorado en Ingeniería Eléctrica, Electrónica y Automática por la Universidad Carlos III de MadridPresidente: Pedro Manuel Urbano de Almeida Lima.- Secretaria: María Dolores Blanco Rojas.- Vocal: Antonio Fernández Caballer
Recommended from our members
AI-based soft module for safe human–robot interaction towards 4D printing
Soft robotic modules have potential use for therapeutic and educational purposes. To do so, they need to be safe, soft, smart, and customizable to serve individuals’ different preferences and personalities. A safe modular robotic product made of soft materials, particularly silicon, programmed by artificial intelligence algorithms and developed via additive manufacturing would be promising. This study focuses on the safe tactile interaction between humans and robots by means of soft material characteristics for translating physical communication to auditory. The embedded vibratory sensors used to stimulate touch senses transmitted through soft materials are presented. The soft module was developed and verified successfully to react to three different patterns of human–robot contact, particularly users’ touches, and then communicate the type of contact with sound. The study develops and verifies a model that can classify different tactile gestures via machine learning algorithms for safe human–robot physical interaction. The system accurately recognizes the gestures and shapes of three-dimensional (3D) printed soft modules. The gestures used for the experiment are the three most common, including slapping, squeezing, and tickling. The model builds on the concept of how safe human–robot physical interactions could help with cognitive and behavioral communication. In this context, the ability to measure, classify, and reflect the behavior of soft materials in robotic modules represents a prerequisite for endowing robotic materials in additive manufacturing for safe interaction with humans
Asynchronous federated learning system for human-robot touch interaction
Artificial intelligence and robotics are advancing at an incredible pace; however, there is a risk associated with the data privacy and personal information of users interacting with these systems and platforms. In this context, the federated learning approach emerged to enable large-scale, distributed learning without the need to transmit or store any information necessary to train the learning models. In a previous paper, we presented a system capable of detecting, locating, and classifying what kind of contact occurs between humans and one of our robots using innovative contact microphone technology. In this work we go further, improving the previously presented touch system with a multi-user, multi-robot, distributed, and scalable learning approach that is able to learn in a collaborative and incremental way while respecting the privacy of the user's information. The system has been successfully evaluated in a real environment with 28 different users divided in 7 different groups. To assess the performance of our system with this federated learning approach, we compared it to the same distributed learning system without federated learning. That is, the control group for this comparison is a central node directly receiving all the training examples obtained by each robot locally. We found that in this context the inclusion of federated learning improves the results concerning traditional distributed learning.The research leading to these results has received funding from the projects: Robots Sociales para Estimulación Física, Cognitiva y Afectiva de Mayores (ROSES), RTI2018-096338-B-I00, funded by the Ministerio de Ciencia, Innovación y Universidades; Robots sociales para mitigar la soledad y el aislamiento en mayores (SOROLI), PID2021-123941OA-I00, funded by Agencia Estatal de Investigación (AEI), Spanish Ministerio de Ciencia e Innovación; the project PLEC2021-007819, funded by MCIN/AEI/10.13039/501100011033 and by the European Union NextGenerationEU/PRTR, and RoboCity2030-DIH-CM, Madrid Robotics Digital Innovation Hub, S2018/NMT-4331, funded by Programas de Actividades I+D en la Comunidad de Madrid and cofunded by the European Social Funds (FSE) of the EU. Funding for APC: Universidad Carlos III de Madrid (Read & Publish Agreement CRUE-CSIC 2022 )
Machine Learning in Sensors and Imaging
Machine learning is extending its applications in various fields, such as image processing, the Internet of Things, user interface, big data, manufacturing, management, etc. As data are required to build machine learning networks, sensors are one of the most important technologies. In addition, machine learning networks can contribute to the improvement in sensor performance and the creation of new sensor applications. This Special Issue addresses all types of machine learning applications related to sensors and imaging. It covers computer vision-based control, activity recognition, fuzzy label classification, failure classification, motor temperature estimation, the camera calibration of intelligent vehicles, error detection, color prior model, compressive sensing, wildfire risk assessment, shelf auditing, forest-growing stem volume estimation, road management, image denoising, and touchscreens
Sensing with Earables: A Systematic Literature Review and Taxonomy of Phenomena
Earables have emerged as a unique platform for ubiquitous computing by augmenting ear-worn devices with state-of-the-art sensing. This new platform has spurred a wealth of new research exploring what can be detected on a wearable, small form factor. As a sensing platform, the ears are less susceptible to motion artifacts and are located in close proximity to a number of important anatomical structures including the brain, blood vessels, and facial muscles which reveal a wealth of information. They can be easily reached by the hands and the ear canal itself is affected by mouth, face, and head movements. We have conducted a systematic literature review of 271 earable publications from the ACM and IEEE libraries. These were synthesized into an open-ended taxonomy of 47 different phenomena that can be sensed in, on, or around the ear. Through analysis, we identify 13 fundamental phenomena from which all other phenomena can be derived, and discuss the different sensors and sensing principles used to detect them. We comprehensively review the phenomena in four main areas of (i) physiological monitoring and health, (ii) movement and activity, (iii) interaction, and (iv) authentication and identification. This breadth highlights the potential that earables have to offer as a ubiquitous, general-purpose platform
- …