6,102 research outputs found
Attention and Anticipation in Fast Visual-Inertial Navigation
We study a Visual-Inertial Navigation (VIN) problem in which a robot needs to
estimate its state using an on-board camera and an inertial sensor, without any
prior knowledge of the external environment. We consider the case in which the
robot can allocate limited resources to VIN, due to tight computational
constraints. Therefore, we answer the following question: under limited
resources, what are the most relevant visual cues to maximize the performance
of visual-inertial navigation? Our approach has four key ingredients. First, it
is task-driven, in that the selection of the visual cues is guided by a metric
quantifying the VIN performance. Second, it exploits the notion of
anticipation, since it uses a simplified model for forward-simulation of robot
dynamics, predicting the utility of a set of visual cues over a future time
horizon. Third, it is efficient and easy to implement, since it leads to a
greedy algorithm for the selection of the most relevant visual cues. Fourth, it
provides formal performance guarantees: we leverage submodularity to prove that
the greedy selection cannot be far from the optimal (combinatorial) selection.
Simulations and real experiments on agile drones show that our approach ensures
state-of-the-art VIN performance while maintaining a lean processing time. In
the easy scenarios, our approach outperforms appearance-based feature selection
in terms of localization errors. In the most challenging scenarios, it enables
accurate visual-inertial navigation while appearance-based feature selection
fails to track robot's motion during aggressive maneuvers.Comment: 20 pages, 7 figures, 2 table
Single camera pose estimation using Bayesian filtering and Kinect motion priors
Traditional approaches to upper body pose estimation using monocular vision
rely on complex body models and a large variety of geometric constraints. We
argue that this is not ideal and somewhat inelegant as it results in large
processing burdens, and instead attempt to incorporate these constraints
through priors obtained directly from training data. A prior distribution
covering the probability of a human pose occurring is used to incorporate
likely human poses. This distribution is obtained offline, by fitting a
Gaussian mixture model to a large dataset of recorded human body poses, tracked
using a Kinect sensor. We combine this prior information with a random walk
transition model to obtain an upper body model, suitable for use within a
recursive Bayesian filtering framework. Our model can be viewed as a mixture of
discrete Ornstein-Uhlenbeck processes, in that states behave as random walks,
but drift towards a set of typically observed poses. This model is combined
with measurements of the human head and hand positions, using recursive
Bayesian estimation to incorporate temporal information. Measurements are
obtained using face detection and a simple skin colour hand detector, trained
using the detected face. The suggested model is designed with analytical
tractability in mind and we show that the pose tracking can be
Rao-Blackwellised using the mixture Kalman filter, allowing for computational
efficiency while still incorporating bio-mechanical properties of the upper
body. In addition, the use of the proposed upper body model allows reliable
three-dimensional pose estimates to be obtained indirectly for a number of
joints that are often difficult to detect using traditional object recognition
strategies. Comparisons with Kinect sensor results and the state of the art in
2D pose estimation highlight the efficacy of the proposed approach.Comment: 25 pages, Technical report, related to Burke and Lasenby, AMDO 2014
conference paper. Code sample: https://github.com/mgb45/SignerBodyPose Video:
https://www.youtube.com/watch?v=dJMTSo7-uF
Computational intelligence approaches to robotics, automation, and control [Volume guest editors]
No abstract available
Contour Tracking Control for Mobile Robots applicable to Large-scale Assembly and Additive Manufacturing in Construction
In the construction industry, as well as during the assembly of large-scale components, the required workspaces usually cannot be served by a stationary robot. Instead, mobile robots are used to increase the accessible space. Here, the problem arises that the accuracy of such systems is not sufficient to meet the tolerance requirements of the components to be produced. Furthermore, there is an additional difficulty in the trajectory planning process since the exact dimensions of the pre-manufactured parts are unknown. Hence, existing static planning methods cannot be exerted on every application. Recent approaches present dynamic planning algorithms based on specific component characteristics. For example, the latest methods follow the contour by a force-controlled motion or detect features with a camera. However, in several applications such as welding or additive manufacturing in construction, no contact force is generated that could be controlled. Vision-based approaches are generally restricted by varying materials and lighting conditions, often found in large-scale construction. For these reasons, we propose a more robust approach without measuring contact forces, which, for example, applies to large-scale additive manufacturing. We based our algorithm on a high-precision 2D line laser, capable of detecting different feature contours regardless of material or lightning. The laser is mounted to the robot's end-effector and provides a depth profile of the component's surface. From this depth data, we determine the target contour and control the manipulator to follow it. Simultaneously we vary the robot's speed to adjust the feed rate depending on the contour's shape, maintaining a constant material application rate. As a proof of concept, we apply the algorithm to the additive manufacturing of two-layer linear structures made from spray PU foam. When making these structures, each layer must be positioned precisely on the previous layer to obtain a straight wall and prevent elastic buckling or plastic collapse. Initial experiments show improved layer alignment within 10 % of the layer width, as well as better layer height consistency and process reliability
A one decade survey of autonomous mobile robot systems
Recently, autonomous mobile robots have gained popularity in the modern world due to their relevance technology and application in real world situations. The global market for mobile robots will grow significantly over the next 20 years. Autonomous mobile robots are found in many fields including institutions, industry, business, hospitals, agriculture as well as private households for the purpose of improving day-to-day activities and services. The development of technology has increased in the requirements for mobile robots because of the services and tasks provided by them, like rescue and research operations, surveillance, carry heavy objects and so on. Researchers have conducted many works on the importance of robots, their uses, and problems. This article aims to analyze the control system of mobile robots and the way robots have the ability of moving in real-world to achieve their goals. It should be noted that there are several technological directions in a mobile robot industry. It must be observed and integrated so that the robot functions properly: Navigation systems, localization systems, detection systems (sensors) along with motion and kinematics and dynamics systems. All such systems should be united through a control unit; thus, the mission or work of mobile robots are conducted with reliability
A reconfigurable hybrid intelligent system for robot navigation
Soft computing has come of age to o er us a wide array of powerful and e cient algorithms
that independently matured and in
uenced our approach to solving problems in robotics,
search and optimisation. The steady progress of technology, however, induced a
ux of new
real-world applications that demand for more robust and adaptive computational paradigms,
tailored speci cally for the problem domain. This gave rise to hybrid intelligent systems, and
to name a few of the successful ones, we have the integration of fuzzy logic, genetic algorithms
and neural networks. As noted in the literature, they are signi cantly more powerful than
individual algorithms, and therefore have been the subject of research activities in the past
decades. There are problems, however, that have not succumbed to traditional hybridisation
approaches, pushing the limits of current intelligent systems design, questioning their solutions
of a guarantee of optimality, real-time execution and self-calibration. This work presents an
improved hybrid solution to the problem of integrated dynamic target pursuit and obstacle
avoidance, comprising of a cascade of fuzzy logic systems, genetic algorithm, the A* search
algorithm and the Voronoi diagram generation algorithm
- …