9,445 research outputs found
Rearrange Indoor Scenes for Human-Robot Co-Activity
We present an optimization-based framework for rearranging indoor furniture
to accommodate human-robot co-activities better. The rearrangement aims to
afford sufficient accessible space for robot activities without compromising
everyday human activities. To retain human activities, our algorithm preserves
the functional relations among furniture by integrating spatial and semantic
co-occurrence extracted from SUNCG and ConceptNet, respectively. By defining
the robot's accessible space by the amount of open space it can traverse and
the number of objects it can reach, we formulate the rearrangement for
human-robot co-activity as an optimization problem, solved by adaptive
simulated annealing (ASA) and covariance matrix adaptation evolution strategy
(CMA-ES). Our experiments on the SUNCG dataset quantitatively show that
rearranged scenes provide an average of 14% more accessible space and 30% more
objects to interact with. The quality of the rearranged scenes is qualitatively
validated by a human study, indicating the efficacy of the proposed strategy.Comment: 7 pages, 7 figures; Accepted by ICRA 202
Language-guided Active Sensing of Confined, Cluttered Environments via Object Rearrangement Planning
Language-guided active sensing is a robotics subtask where a robot with an
onboard sensor interacts efficiently with the environment via object
manipulation to maximize perceptual information, following given language
instructions. These tasks appear in various practical robotics applications,
such as household service, search and rescue, and environment monitoring.
Despite many applications, the existing works do not account for language
instructions and have mainly focused on surface sensing, i.e., perceiving the
environment from the outside without rearranging it for dense sensing.
Therefore, in this paper, we introduce the first language-guided active sensing
approach that allows users to observe specific parts of the environment via
object manipulation. Our method spatially associates the environment with
language instructions, determines the best camera viewpoints for perception,
and then iteratively selects and relocates the best view-blocking objects to
provide the dense perception of the region of interest. We evaluate our method
against different baseline algorithms in simulation and also demonstrate it in
real-world confined cabinet-like settings with multiple unknown objects. Our
results show that the proposed method exhibits better performance across
different metrics and successfully generalizes to real-world complex scenarios.Comment: Accepted in IEEE/RAS ICRA'2
Automated sequence and motion planning for robotic spatial extrusion of 3D trusses
While robotic spatial extrusion has demonstrated a new and efficient means to
fabricate 3D truss structures in architectural scale, a major challenge remains
in automatically planning extrusion sequence and robotic motion for trusses
with unconstrained topologies. This paper presents the first attempt in the
field to rigorously formulate the extrusion sequence and motion planning (SAMP)
problem, using a CSP encoding. Furthermore, this research proposes a new
hierarchical planning framework to solve the extrusion SAMP problems that
usually have a long planning horizon and 3D configuration complexity. By
decoupling sequence and motion planning, the planning framework is able to
efficiently solve the extrusion sequence, end-effector poses, joint
configurations, and transition trajectories for spatial trusses with
nonstandard topologies. This paper also presents the first detailed computation
data to reveal the runtime bottleneck on solving SAMP problems, which provides
insight and comparing baseline for future algorithmic development. Together
with the algorithmic results, this paper also presents an open-source and
modularized software implementation called Choreo that is machine-agnostic. To
demonstrate the power of this algorithmic framework, three case studies,
including real fabrication and simulation results, are presented.Comment: 24 pages, 16 figure
Pose-Based Tactile Servoing: Controlled Soft Touch using Deep Learning
This article describes a new way of controlling robots using soft tactile
sensors: pose-based tactile servo (PBTS) control. The basic idea is to embed a
tactile perception model for estimating the sensor pose within a servo control
loop that is applied to local object features such as edges and surfaces. PBTS
control is implemented with a soft curved optical tactile sensor (the BRL
TacTip) using a convolutional neural network trained to be insensitive to
shear. In consequence, robust and accurate controlled motion over various
complex 3D objects is attained. First, we review tactile servoing and its
relation to visual servoing, before formalising PBTS control. Then, we assess
tactile servoing over a range of regular and irregular objects. Finally, we
reflect on the relation to visual servo control and discuss how controlled soft
touch gives a route towards human-like dexterity in robots.Comment: A summary video is available here https://youtu.be/12-DJeRcfn0 *NL
and JL contributed equally to this wor
Recent Advancements in Augmented Reality for Robotic Applications: A Survey
Robots are expanding from industrial applications to daily life, in areas such as medical robotics, rehabilitative robotics, social robotics, and mobile/aerial robotics systems. In recent years, augmented reality (AR) has been integrated into many robotic applications, including medical, industrial, human–robot interactions, and collaboration scenarios. In this work, AR for both medical and industrial robot applications is reviewed and summarized. For medical robot applications, we investigated the integration of AR in (1) preoperative and surgical task planning; (2) image-guided robotic surgery; (3) surgical training and simulation; and (4) telesurgery. AR for industrial scenarios is reviewed in (1) human–robot interactions and collaborations; (2) path planning and task allocation; (3) training and simulation; and (4) teleoperation control/assistance. In addition, the limitations and challenges are discussed. Overall, this article serves as a valuable resource for working in the field of AR and robotic research, offering insights into the recent state of the art and prospects for improvement
Mobile Manipulation Hackathon: Moving into Real World Applications
The Mobile Manipulation Hackathon was held in late 2018 during the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) to showcase the latest applications of wheeled robotic manipulators. The challenge had an open format, where teams developed an application using simulation tools and integrated it into a robotic platform. This article presents the competition and analyzes the results, with information gathered during the event and from a survey circulated among the finalist teams. We provide an overview of the mobile manipulation field, identify key areas required for further development to facilitate the implementation of mobile manipulators in real applications, and discuss ideas about how to structure future hackathon-style competitions to enhance their impact on the scientific and industrial communities.Peer ReviewedPostprint (published version
Scalable Approach to Uncertainty Quantification and Robust Design of Interconnected Dynamical Systems
Development of robust dynamical systems and networks such as autonomous
aircraft systems capable of accomplishing complex missions faces challenges due
to the dynamically evolving uncertainties coming from model uncertainties,
necessity to operate in a hostile cluttered urban environment, and the
distributed and dynamic nature of the communication and computation resources.
Model-based robust design is difficult because of the complexity of the hybrid
dynamic models including continuous vehicle dynamics, the discrete models of
computations and communications, and the size of the problem. We will overview
recent advances in methodology and tools to model, analyze, and design robust
autonomous aerospace systems operating in uncertain environment, with stress on
efficient uncertainty quantification and robust design using the case studies
of the mission including model-based target tracking and search, and trajectory
planning in uncertain urban environment. To show that the methodology is
generally applicable to uncertain dynamical systems, we will also show examples
of application of the new methods to efficient uncertainty quantification of
energy usage in buildings, and stability assessment of interconnected power
networks
- …