74 research outputs found
Vision-based trajectory control of unsensored robots to increase functionality, without robot hardware modication
In nuclear decommissioning operations, very rugged remote manipulators are used, which lack proprioceptive joint angle sensors. Hence these machines are simply tele-operated, where a human operator controls each joint of the robot individually using a teach pendant or a set of switches. Moreover, decommissioning tasks often involve forceful interactions between the environment and powerful tools at the robot's end-effector. Such interactions can result in complex dynamics, large torques at the robot's joints, and can also lead to erratic movements of a mobile manipulator's base frame with respect to the task space. This Thesis seeks to address these problems by, firstly, showing how
the configuration of such robots can be tracked in real-time by a vision system and fed back into a trajectory control scheme. Secondly, the Thesis investigates the dynamics of robot-environment contacts, and proposes several control schemes for detecting, coping
with, and also exploiting such contacts. Several contributions are advanced in this Thesis. Specifically a control framework is presented which exploits the constraints arising at contact points to effectively reduce commanded torques to perform tasks; methods are advanced to estimate the constraints arising from contacts in a number of situations, using only kinematic quantities; a framework is proposed to estimate the
configuration of a manipulator using a single monocular camera; and finally, a general control framework is described which uses all of the above contributions to servo a manipulator. The results of a number of experiments are presented which demonstrate the feasibility of the proposed methods
Angioleiomyoma of the knee: An uncommon cause of leg pain. A systematic review of the literature
Abstract Objective Angioleiomyoma is a rare benign painful soft tissue tumor, whose knee location is rare. Due its rarity, and not characteristic aspect on MRI the preoperative diagnosis is difficult. Methods We performed a systematic review of the literature, including a case of venous type angioleiomyoma that we have recently managed. Results A total of 24 published papers with 30 cases (including our illustrative case) were identified and included in our review. The mean patient age was 42.3 years (range18-63). The average size of the lesion was 17.8 mm. The presenting symptom was leg pain in 90% of cases. On magnetic resonance imaging (MRI), the lesion appeared isointense in T1 in 80% of cases and hyperintense on T2 in 90% of cases. Avid homogeneous enhancement after gadolinium administration was detected in 94% of cases. All patients underwent surgery and total resection was achieved in 100% of cases. No recurrence was observed after a mean follow-up of 19.5 months. Conclusion Angioleiomyoma occurs rarely in the knee and generally is associated with localized or radiating pain. The preoperative diagnosis is difficult also after completion of MRI study and requires high index of suspicion. Angioleiomyoma widens the spectrum of soft tissue lesions of the extremities and should be included in the differential diagnosis of lesions in this area
Zero-Shot Multi-Modal Artist-Controlled Retrieval and Exploration of 3D Object Sets
When creating 3D content, highly specialized skills are generally needed to
design and generate models of objects and other assets by hand. We address this
problem through high-quality 3D asset retrieval from multi-modal inputs,
including 2D sketches, images and text. We use CLIP as it provides a bridge to
higher-level latent features. We use these features to perform a multi-modality
fusion to address the lack of artistic control that affects common data-driven
approaches. Our approach allows for multi-modal conditional feature-driven
retrieval through a 3D asset database, by utilizing a combination of input
latent embeddings. We explore the effects of different combinations of feature
embeddings across different input types and weighting methods
Object Handovers: a Review for Robotics
This article surveys the literature on human-robot object handovers. A
handover is a collaborative joint action where an agent, the giver, gives an
object to another agent, the receiver. The physical exchange starts when the
receiver first contacts the object held by the giver and ends when the giver
fully releases the object to the receiver. However, important cognitive and
physical processes begin before the physical exchange, including initiating
implicit agreement with respect to the location and timing of the exchange.
From this perspective, we structure our review into the two main phases
delimited by the aforementioned events: 1) a pre-handover phase, and 2) the
physical exchange. We focus our analysis on the two actors (giver and receiver)
and report the state of the art of robotic givers (robot-to-human handovers)
and the robotic receivers (human-to-robot handovers). We report a comprehensive
list of qualitative and quantitative metrics commonly used to assess the
interaction. While focusing our review on the cognitive level (e.g.,
prediction, perception, motion planning, learning) and the physical level
(e.g., motion, grasping, grip release) of the handover, we briefly discuss also
the concepts of safety, social context, and ergonomics. We compare the
behaviours displayed during human-to-human handovers to the state of the art of
robotic assistants, and identify the major areas of improvement for robotic
assistants to reach performance comparable to human interactions. Finally, we
propose a minimal set of metrics that should be used in order to enable a fair
comparison among the approaches.Comment: Review paper, 19 page
Planning maximum-manipulability cutting paths
This paper presents a method for constrained motion planning from vision, which enables a robot to move its end-effector over an observed surface, given start and destination points. The robot has no prior knowledge of the surface shape but observes it from a noisy point cloud. We consider the multi-objective optimisation problem of finding robot trajectories which maximise the robot’s manipulability throughout the motion, while also minimising surface-distance travelled between the two points. This work has application in industrial problems of rough robotic cutting, e.g., demolition of the legacy nuclear plant, where the cut path needs not be precise as long as it achieves dismantling. We show how detours in the path can be leveraged to increase the manipulability of the robot at all points along the path. This helps to avoid singularities while maximising the robot’s capability to make small deviations during task execution. We show how a sampling-based planner can be projected onto the Riemannian manifold of a curved surface, and extended to include a term which maximises manipulability. We present the results of empirical experiments, with both simulated and real robots, which are tasked with moving over a variety of different surface shapes. Our planner enables successful task completion while ensuring significantly greater manipulability when compared against a conventional RRT* planner
Towards Advanced Robotic Manipulations for Nuclear Decommissioning
Despite enormous remote handling requirements, remarkably very few robots are being used by the nuclear industry. Most of the remote handling tasks are still performed manually, using conventional mechanical master‐slave devices. The few robotic manipulators deployed are directly tele‐operated in rudimentary ways, with almost no autonomy or even a pre‐programmed motion. In addition, majority of these robots are under‐sensored (i.e. with no proprioception), which prevents them to use for automatic tasks. In this context, primarily this chapter discusses the human operator performance in accomplishing heavy‐duty remote handling tasks in hazardous environments such as nuclear decommissioning. Multiple factors are evaluated to analyse the human operators’ performance and workload. Also, direct human tele‐operation is compared against human‐supervised semi‐autonomous control exploiting computer vision. Secondarily, a vision‐guided solution towards enabling advanced control and automating the under‐sensored robots is presented. Maintaining the coherence with real nuclear scenario, the experiments are conducted in the lab environment and results are discussed
Model-free vision-based shaping of deformable plastic materials
We address the problem of shaping deformable plastic materials using
non-prehensile actions. Shaping plastic objects is challenging, since they are
difficult to model and to track visually. We study this problem, by using
kinetic sand, a plastic toy material which mimics the physical properties of
wet sand. Inspired by a pilot study where humans shape kinetic sand, we define
two types of actions: \textit{pushing} the material from the sides and
\textit{tapping} from above. The chosen actions are executed with a robotic arm
using image-based visual servoing. From the current and desired view of the
material, we define states based on visual features such as the outer contour
shape and the pixel luminosity values. These are mapped to actions, which are
repeated iteratively to reduce the image error until convergence is reached.
For pushing, we propose three methods for mapping the visual state to an
action. These include heuristic methods and a neural network, trained from
human actions. We show that it is possible to obtain simple shapes with the
kinetic sand, without explicitly modeling the material. Our approach is limited
in the types of shapes it can achieve. A richer set of action types and
multi-step reasoning is needed to achieve more sophisticated shapes.Comment: Accepted to The International Journal of Robotics Research (IJRR
Vision-Guided State Estimation and Control of Robotic Manipulators Which Lack Proprioceptive Sensors
This paper presents a vision-based approach for estimating the configuration of, and providing control signals for, an under-sensored robot manipulator using a single monocular camera. Some remote manipulators, used for decommissioning tasks in the nuclear industry, lack proprioceptive sensors because electronics are vulnerable to radiation. Additionally, even if proprioceptive joint sensors could be retrofitted, such heavy-duty manipulators are often deployed on mobile vehicle platforms, which are significantly and erratically perturbed when powerful hydraulic drilling or cutting tools are deployed at the end-effector. In these scenarios, it would be beneficial to use external sensory information, e.g. vision, for estimating the robot configuration with respect to the scene or task. Conventional visual servoing methods typically rely on joint encoder values for controlling the robot. In contrast, our framework assumes that no joint encoders are available, and estimates the robot configuration by visually tracking several parts of the robot, and then enforcing equality between a set of transformation matrices which relate the frames of the camera, world and tracked robot parts. To accomplish this, we propose two alternative methods based on optimisation. We evaluate the performance of our developed framework by visually tracking the pose of a conventional robot arm, where the joint encoders are used to provide ground-truth for evaluating the precision of the vision system. Additionally, we evaluate the precision with which visual feedback can be used to control the robot's end-effector to follow a desired trajectory
Model Estimation and Control of Compliant Contact Normal Force
This paper proposes a method to realize desired contact normal forces between humanoids and their compliant environment. By using contact models, desired contact forces are converted to desired deformations of compliant surfaces. To achieve desired forces, deformations are controlled by controlling the contact point positions. Parameters of contact models are assumed to be known or estimated using the approach described in this paper. The proposed methods for estimating the contact parameters and controlling the contact normal force are implemented on a LWR KUKA IV arm. To verify both methods, experiments are performed with the KUKA arm while its end-effector is in contact with two different soft objects
- …