104 research outputs found

    Enhancing Supermarket Robot Interaction: A Multi-Level LLM Conversational Interface for Handling Diverse Customer Intents

    Full text link
    This paper presents the design and evaluation of a novel multi-level LLM interface for supermarket robots to assist customers. The proposed interface allows customers to convey their needs through both generic and specific queries. While state-of-the-art systems like OpenAI's GPTs are highly adaptable and easy to build and deploy, they still face challenges such as increased response times and limitations in strategic control of the underlying model for tailored use-case and cost optimization. Driven by the goal of developing faster and more efficient conversational agents, this paper advocates for using multiple smaller, specialized LLMs fine-tuned to handle different user queries based on their specificity and user intent. We compare this approach to a specialized GPT model powered by GPT-4 Turbo, using the Artificial Social Agent Questionnaire (ASAQ) and qualitative participant feedback in a counterbalanced within-subjects experiment. Our findings show that our multi-LLM chatbot architecture outperformed the benchmarked GPT model across all 13 measured criteria, with statistically significant improvements in four key areas: performance, user satisfaction, user-agent partnership, and self-image enhancement. The paper also presents a method for supermarket robot navigation by mapping the final chatbot response to correct shelf numbers, enabling the robot to sequentially navigate towards the respective products, after which lower-level robot perception, control, and planning can be used for automated object retrieval. We hope this work encourages more efforts into using multiple, specialized smaller models instead of relying on a single powerful, but more expensive and slower model

    Enhancing supermarket robot interaction: an equitable multi-level LLM conversational interface for handling diverse customer intents

    Get PDF
    This paper presents the design and evaluation of a comprehensive system to develop voice-based interfaces to support users in supermarkets. These interfaces enable shoppers to convey their needs through both generic and specific queries. Although customisable state-of-the-art systems like GPTs from OpenAI are easily accessible and adaptable, featuring low-code deployment with options for functional integration, they still face challenges such as increased response times and limitations in strategic control for tailored use cases and cost optimization. Motivated by the goal of crafting equitable and efficient conversational agents with a touch of personalisation, this study advances on two fronts: 1) a comparative analysis of four popular off-the-shelf speech recognition technologies to identify the most accurate model for different genders (male/female) and languages (English/Dutch) and 2) the development and evaluation of a novel multi-LLM supermarket chatbot framework, comparing its performance with a specialized GPT model powered by the GPT-4 Turbo, using the Artificial Social Agent Questionnaire (ASAQ) and qualitative participant feedback. Our findings reveal that OpenAI’s Whisper leads in speech recognition accuracy between genders and languages and that our proposed multi-LLM chatbot architecture, which outperformed the benchmarked GPT model in performance, user satisfaction, user-agent partnership, and self-image enhancement, achieved statistical significance in these four key areas out of the 13 evaluated aspects that all showed improvements. The paper concludes with a simple method for supermarket robot navigation by mapping the final chatbot response to the correct shelf numbers to which the robot can plan sequential visits. Later, this enables the effective use of low-level perception, motion planning, and control capabilities for product retrieval and collection. We hope that this work encourages more efforts to use multiple specialized smaller models instead of always relying on a single powerful model

    ILoSA: Interactive Learning of Stiffness and Attractors

    Full text link
    Teaching robots how to apply forces according to our preferences is still an open challenge that has to be tackled from multiple engineering perspectives. This paper studies how to learn variable impedance policies where both the Cartesian stiffness and the attractor can be learned from human demonstrations and corrections with a user-friendly interface. The presented framework, named ILoSA, uses Gaussian Processes for policy learning, identifying regions of uncertainty and allowing interactive corrections, stiffness modulation and active disturbance rejection. The experimental evaluation of the framework is carried out on a Franka-Emika Panda in three separate cases with unique force interaction properties: 1) pulling a plug wherein a sudden force discontinuity occurs upon successful removal of the plug, 2) pushing a box where a sustained force is required to keep the robot in motion, and 3) wiping a whiteboard in which the force is applied perpendicular to the direction of movement

    Dynamic movement primitives in robotics:A tutorial survey

    Get PDF
    Biological systems, including human beings, have the innate ability to perform complex tasks in a versatile and agile manner. Researchers in sensorimotor control have aimed to comprehend and formally define this innate characteristic. The idea, supported by several experimental findings, that biological systems are able to combine and adapt basic units of motion into complex tasks finally leads to the formulation of the motor primitives’ theory. In this respect, Dynamic Movement Primitives (DMPs) represent an elegant mathematical formulation of the motor primitives as stable dynamical systems and are well suited to generate motor commands for artificial systems like robots. In the last decades, DMPs have inspired researchers in different robotic fields including imitation and reinforcement learning, optimal control, physical interaction, and human–robot co-working, resulting in a considerable amount of published papers. The goal of this tutorial survey is two-fold. On one side, we present the existing DMP formulations in rigorous mathematical terms and discuss the advantages and limitations of each approach as well as practical implementation details. In the tutorial vein, we also search for existing implementations of presented approaches and release several others. On the other side, we provide a systematic and comprehensive review of existing literature and categorize state-of-the-art work on DMP. The paper concludes with a discussion on the limitations of DMPs and an outline of possible research directions.</p

    Interactive Imitation Learning of Bimanual Movement Primitives

    Get PDF
    Performing bimanual tasks with dual robotic setups can drastically increase the impact on industrial and daily life applications. However, performing a bimanual task brings many challenges, like synchronization and coordination of the single-arm policies. This article proposes the Safe, Interactive Movement Primitives Learning (SIMPLe) algorithm, to teach and correct single or dual arm impedance policies directly from human kinesthetic demonstrations. Moreover, it proposes a novel graph encoding of the policy based on Gaussian Process Regression (GPR) where the single-arm motion is guaranteed to converge close to the trajectory and then towards the demonstrated goal. Regulation of the robot stiffness according to the epistemic uncertainty of the policy allows for easily reshaping the motion with human feedback and/or adapting to external perturbations. We tested the SIMPLe algorithm on a real dual-arm setup where the teacher gave separate single-arm demonstrations and then successfully synchronized them only using kinesthetic feedback or where the original bimanual demonstration was locally reshaped to pick a box at a different height

    Augmenting visual feedback with visualized interaction forces in haptic-assisted virtual-reality teleoperation

    Get PDF
    In recent years, providing additional visual feedback about the interaction forces has been found to offer benefits to haptic-assisted teleoperation. However, there is limited insight into the effects of the design of force feedback-related visual cues and the type of visual display on the performance of teleoperation of robotic arms executing industrial tasks. In this study, we provide new insights into this interaction by extending these findings to the haptic assistance teleoperation of a simulated robotic arm in a virtual environment, in which the haptic assistance is comprised of a set of virtual fixtures. We design a novel method for providing visual cues about the interaction forces to complement the haptic assistance and augment visual feedback in virtual reality with a head-mounted display. We evaluate the visual cues method and head-mounted display method through human factors experiments in a teleoperated dross removal use case. The results show that both methods are beneficial for task performance, each of them having stronger points in different aspects of the operation. The visual cues method was found to significantly improve safety in terms of peak collision force, whereas the head-mounted display additionally improves the performance significantly. Furthermore, positive scores of the subjective analysis indicate an increased user acceptance of both methods. This work provides a new study on the importance of visual feedback related to (interaction) forces and spatial information for haptic assistance and provides two methods to take advantage of its potential benefits in the teleoperation of robotic arms

    Unifying Speed-Accuracy Trade-Off and Cost-Benefit Trade-Off in Human Reaching Movements

    No full text
    Two basic trade-offs interact while our brain decides how to move our body. First, with the cost-benefit trade-off, the brain trades between the importance of moving faster toward a target that is more rewarding and the increased muscular cost resulting from a faster movement. Second, with the speed-accuracy trade-off, the brain trades between how accurate the movement needs to be and the time it takes to achieve such accuracy. So far, these two trade-offs have been well studied in isolation, despite their obvious interdependence. To overcome this limitation, we propose a new model that is able to simultaneously account for both trade-offs. The model assumes that the central nervous system maximizes the expected utility resulting from the potential reward and the cost over the repetition of many movements, taking into account the probability to miss the target. The resulting model is able to account for both the speed-accuracy and the cost-benefit trade-offs. To validate the proposed hypothesis, we confront the properties of the computational model to data from an experimental study where subjects have to reach for targets by performing arm movements in a horizontal plane. The results qualitatively show that the proposed model successfully accounts for both cost-benefit and speed-accuracy trade-offs

    ACCELERATED ROBOTIC LEARNING FOR INTERACTION WITH ENVIRONMENT AND HUMAN BASED ON SENSORY-MOTOR LEARNING

    Full text link
    V prihodnje je premik robotov iz strukturiranih/nadzorovanih industrijskih in laboratorijskih okolij v človekov vsakdanjik neizogiben. Za uspešno vpeljavo robotov v naš vsakdanjik je treba rešiti dva osnovna problema, povezana z izvajanjem takih nalog. (1) Naloge, potrebe, orodja in okolje se razlikujejo med uporabniki. To zahteva pridobitev velike baze zelo specifičnih znanj, zato mora biti prenos znanja z uporabnika na robota enostaven, prilagodljiv in hiter. (2) Pri izvedbi večine vsakdanjih nalog se srečujemo s žičnimi interakcijami z nepredvidljivim in nestrukturiranim okoljem. Zato je pridobivanje modelov okolja zapleten postopek, kar oteži vodenje robota s klasičnimi pristopi. Za dobro delovanje robotov v takem okolju moramo najti alternativne načine učenja žičnih interakcij robota z okoljem. V prvem poglavju je najprej predstavljen uvod v tematiko s pregledom dosedanjega dela na tem področju. Nato so predstavljeni glavni cilji disertacije. Na koncu poglavja pa je povzetek vseh eksperimentov, ki so bili izvedeni v okviru disertacije. V drugem poglavju je najprej splošno predstavljena metoda učenja robotov z vključitvijo človekovega senzorično-motoričnega sistema v robotovo regulacijsko zanko. Ta temelji na sposobnosti človekovega senzorično-motoričnega učenja, ki mu omogoča prilagoditev na vodenje robota in kasnejši prenos pridobljenega znanja. Sledi opis metod za predstavitev strategije vodenja (znanja). Te obsegajo senzorično-motorični pare, trajektorije gibanja in adaptivne oscilatorje za opis stanja periodičnega gibanja. V zadnjem sklopu pa sta predstavljeni dve metodi strojenega učenja, ki sta bili uporabljeni za robotsko učenje pridobljenih strategij vodenja (Gaussov regresijski proces in lokalno utežena regresija). V tretjem poglavju je predstavljena metoda učenja humanoidnih robotov za primerne odzive telesa na žične interakcije s človekom in z okoljem. Pri tem je bila razvita metoda za pretvorbo senzoričnih informacij o stanju robotovega telesa v senzorično vzbujanje človeka. S tem je bila demonstratorju posredovana potrebna povratna informacija o stanju dinamike robotovega telesa med učnim postopkom. V okviru tega je bil razvit poseben haptični vmesnik, ki je izvajal silo na telo demonstratorja. V drugem delu poglavja je predstavljena metoda sprotnega učenja robota na podlagi delitve odgovornosti med trenutno naučeno strategijo vodenja robota in demonstratorjem. Ta omogoča postopen prenos odgovornosti vodenja z demonstratorja na robota in omogoča dodatno povratno informacije o stanju učenja. V zadnjem delu pa je predlagana metoda za združevanje demonstriranih strategij vodenja robotskega telesa z vodenjem gibanja roke na osnovi inverzne kinematike. V četrtem poglavju so predstavljene predlagane metode učenja robotske manipulacije z nestrukturiranim in nepredvidljivim okoljem. Te metode temeljijo na zmožnosti demonstratorja direktne modulacije in učenja impedance robotske roke. V ta namen so bile razvite metode, ki omogočajo demonstratorju vodenje togosti v realnem času. Pri tem smo reševali naloge, povezane z uporabo elementarnih orodij, s sodelovanjem robota z uporabnikom in sestavljanjem predmetov. Te lastnosti so ključne pri prihodnjem delovanju robotov v človekovem vsakdanjiku ali pri njihovi udeležbi pri raziskovanju vesolja, kjer so sredstva omejena. V petem poglavju je predstavljena metoda vodenja eksoskeletov. Ti mehanizmi obdajajo dele človeškega telesa in direktno pomagajo pri gibanju v sklepih. V okviru vpeljave robotov v vsakdanjik eksoskeleti predstavljajo komplement humanoidnim robotom, katerih namen je nuditi pomoč človeku na bolj posrednem nivoju. Predlagana metoda vodenja temelji na minimizaciji uporabnikove mišične aktivnosti prek adaptivnega učenja podpornih sklepnih navorov, ki jih izvaja eksoskelet. Glavna prednost metode je, da ne potrebuje modelov človeka in robota. Potrebni kompenzacijski navori se nenehno prilagajajo trenutnim pogojem. Metodo smo preizkusili z eksperimenti na več subjektih in pri tem analizirali medsebojno prilagajanje eksoskeleta ter uporabnika. V zadnjem poglavju sledi zaključek, v katerem so povzeti glavni prispevki disertacije k znanosti.It is inevitable that the robots will move from the structured and controlled environments, such as industry and laboratories, into our daily lives. To achieve the integration of robots into our daily lives, we have to solve two fundamental problems related to these tasks. (1) Each individual robot user has a different environment, needs, tasks and tools. This demands acquisition of a vast amount of very specific skills, thus the transfer of the skill from the user to the robot must be intuitive, adaptable and fast. (2) Many daily tasks require us to deal with physical interaction with unpredictable and unstructured environment. Because of this, the acquisition of models is very complex and makes the classical robot control difficult. To make robots successfully operate in such environment, it is necessary to find alternative ways to teach the robot how do deal with such interactions. In the first part of the first chapter, there is an introduction into the research field with an overview of the state-of-the-art. The second part explains the goals of the thesis. The last part contains an overview of the performed experiments. The second chapter presents the human-in-the-loop teaching method. The method is based on human sensorimotor learning ability that allows the demonstrator to first obtain the skill necessary for controlling the robot and then transfer that skill to the robot. This is followed by a presentation of methods to encode the control strategy (skill). These methods include: sensorimotor pairs, trajectory of motion and adaptive oscillators for describing the state of periodic motion. In the last part of this chapter, we present two machine learning methods that were utilised in our methods (Gaussian Process Regression and Locally Weighted Regression). The third chapter presents the proposed method for teaching humanoid robots how to deal with physical interaction of its body with human and environment. To this end, a method for converting robot sensory information into a human sensory stimulation was developed to give the demonstrator the necessary feedback about the robot body dynamics during the teaching process. In this scope, we developed a special haptic interface that exerted forces on the demonstrator\u27s body. After that, we present a method for on-line robot learning where the control of the robot body is shared between the currently learnt strategy and human demonstrator. This enables a gradual transfer of the control responsibility from the demonstrator to the robot and offers an additional feedback about the state of the learning. In the last part, we propose a method that allows merging human-demonstrated posture-control skill with arm motion control based on inverse kinematics solution. The fourth chapter presents the proposed methods for teaching robot how to manipulate with unstructured and unpredictable environment. These methods were based on the ability of demonstrator to modulate and teach the impedance of robotic arm. We developed methods that allow the demonstrator to control the robot\u27s stiffness in real-time. This approach was then used to solve tasks related to use of elementary tools, human-robot cooperation and part assembly. These tasks are crucial for future robot operation in human daily lives or in their participation in space exploration, where the available means are limited. The fifth chapter presents a method for exoskeleton control. These devices are made to enclose the human body parts and directly assist the motion in the joints. In the framework of integration of robots into the human daily lives, exoskeletons are a complement to the humanoid robots that are designed to provide assistance on a more indirect level. The proposed control method is based on minimisation of human muscle activity through adaptive learning of robot assistive joint torques. The main advantage of this method is that it does not require models of human and robot. Necessary compensation torques are adaptively derived according to the current conditions. The method was validated on multiple subjects and we analysed the human-robot co-adaptation. The last chapter recapitulates the main contributions of the dissertation and presents its conclusions
    corecore