39 research outputs found

    Tactile perception in hydrogel-based robotic skins using data-driven electrical impedance tomography

    Get PDF
    Combining functional soft materials with electrical impedance tomography is a promising method for developing continuum sensorized soft robotic skins with high resolutions. However, reconstructing the tactile stimuli from surface electrode measurements is a challenging ill-posed modelling problem, with FEM and analytic models facing a reality gap. To counter this, we propose and demonstrate a model-free superposition method which uses small amounts of real-world data to develop deformation maps of a soft robotic skin made from a self-healing ionically conductive hydrogel, the properties of which are affected by temperature, humidity, and damage. We demonstrate how this method outperforms a traditional neural network for small datasets, obtaining an average resolution of 12.1 mm over a 170 mm circular skin. Additionally, we explore how this resolution varies over a series of 15,000 consecutive presses, during which damages are continuously propagated. Finally, we demonstrate applications for functional robotic skins: damage detection/localization, environmental monitoring, and multi-touch recognition - all using the same sensing material

    Multi-modal perception for soft robotic interactions using generative models

    Full text link
    Perception is essential for the active interaction of physical agents with the external environment. The integration of multiple sensory modalities, such as touch and vision, enhances this perceptual process, creating a more comprehensive and robust understanding of the world. Such fusion is particularly useful for highly deformable bodies such as soft robots. Developing a compact, yet comprehensive state representation from multi-sensory inputs can pave the way for the development of complex control strategies. This paper introduces a perception model that harmonizes data from diverse modalities to build a holistic state representation and assimilate essential information. The model relies on the causality between sensory input and robotic actions, employing a generative model to efficiently compress fused information and predict the next observation. We present, for the first time, a study on how touch can be predicted from vision and proprioception on soft robots, the importance of the cross-modal generation and why this is essential for soft robotic interactions in unstructured environments.Comment: Accepted for presentation at IEEE RoboSoft 202

    A bistable soft gripper with mechanically embedded sensing and actuation for fast closed-loop grasping

    Get PDF
    Soft robotic grippers are shown to be high effective for grasping unstructured objects with simple sensing and control strategies. However, they are still limited by their speed, sensing capabilities and actuation mechanism. Hence, their usage have been restricted in highly dynamic grasping tasks. This paper presents a soft robotic gripper with tunable bistable properties for sensor-less dynamic grasping. The bistable mechanism allows us to store arbitrarily large strain energy in the soft system which is then released upon contact. The mechanism also provides flexibility on the type of actuation mechanism as the grasping and sensing phase is completely passive. Theoretical background behind the mechanism is presented with finite element analysis to provide insights into design parameters. Finally, we experimentally demonstrate sensor-less dynamic grasping of an unknown object within 0.02 seconds, including the time to sense and actuate

    Visuo-dynamic self-modelling of soft robotic systems

    Get PDF
    Soft robots exhibit complex nonlinear dynamics with large degrees of freedom, making their modelling and control challenging. Typically, reduced-order models in time or space are used in addressing these challenges, but the resulting simplification limits soft robot control accuracy and restricts their range of motion. In this work, we introduce an end-to-end learning-based approach for fully dynamic modelling of any general robotic system that does not rely on predefined structures, learning dynamic models of the robot directly in the visual space. The generated models possess identical dimensionality to the observation space, resulting in models whose complexity is determined by the sensory system without explicitly decomposing the problem. To validate the effectiveness of our proposed method, we apply it to a fully soft robotic manipulator, and we demonstrate its applicability in controller development through an open-loop optimization-based controller. We achieve a wide range of dynamic control tasks including shape control, trajectory tracking and obstacle avoidance using a model derived from just 90 min of real-world data. Our work thus far provides the most comprehensive strategy for controlling a general soft robotic system, without constraints on the shape, properties, or dimensionality of the system

    Soft Self-Healing Fluidic Tactile Sensors with Damage Detection and Localization Abilities

    Get PDF
    Self-healing sensors have the potential to increase the lifespan of existing sensing technologies, especially in soft robotic and wearable applications. Furthermore, they could bestow additional functionality to the sensing system because of their self-healing ability. This paper presents the design for a self-healing sensor that can be used for damage detection and localization in a continuous manner. The soft sensor can recover full functionality almost instantaneously at room temperature, making the healing process fully autonomous. The working principle of the sensor is based on the measurement of air pressure inside enclosed chambers, making the fabrication and the modeling of the sensors easy. We characterize the force sensing abilities of the proposed sensor and perform damage detection and localization over a one-dimensional and two-dimensional surface using multilateration techniques. The proposed solution is highly scalable, easy-to-build, cheap and even applicable for multi-damage detection.This work was supported by the SHERO project, a Future and Emerging Technologies284(FET) programme of the European Commission (grant agreement ID 828818

    Learning-Based Control Strategies for Soft Robots: Theory, Achievements, and Future Challenges

    Get PDF
    In the last few decades, soft robotics technologies have challenged conventional approaches by introducing new, compliant bodies to the world of rigid robots. These technologies and systems may enable a wide range of applications, including human-robot interaction and dealing with complex environments. Soft bodies can adapt their shape to contact surfaces, distribute stress over a larger area, and increase the contact surface area, thus reducing impact forces

    Learning to stop: a unifying principle for legged locomotion in varying environments.

    Get PDF
    Evolutionary studies have unequivocally proven the transition of living organisms from water to land. Consequently, it can be deduced that locomotion strategies must have evolved from one environment to the other. However, the mechanism by which this transition happened and its implications on bio-mechanical studies and robotics research have not been explored in detail. This paper presents a unifying control strategy for locomotion in varying environments based on the principle of 'learning to stop'. Using a common reinforcement learning framework, deep deterministic policy gradient, we show that our proposed learning strategy facilitates a fast and safe methodology for transferring learned controllers from the facile water environment to the harsh land environment. Our results not only propose a plausible mechanism for safe and quick transition of locomotion strategies from a water to land environment but also provide a novel alternative for safer and faster training of robots
    corecore