543 research outputs found

    Wall-Corner Classification Using Sonar: A New Approach Based on Geometric Features

    Get PDF
    Ultrasonic signals coming from rotary sonar sensors in a robot gives us several features about the environment. This enables us to locate and classify the objects in the scenario of the robot. Each object and reflector produces a series of peaks in the amplitude of the signal. The radial and angular position of the sonar sensor gives information about location and their amplitudes offer information about the nature of the surface. Early works showed that the amplitude can be modeled and used to classify objects with very good results at short distances—80% average success in classifying both walls and corners at distances less than 1.5 m. In this paper, a new set of geometric features derived from the amplitude analysis of the echo is presented. These features constitute a set of characteristics that can be used to improve the results of classification at distances from 1.5 m to 4 m. Also, a comparative study on classification algorithms widely used in pattern recognition techniques has been carried out for sensor distances ranging between 0.5 to 4 m, and with incidence angles ranging between 20° to 70°. Experimental results show an enhancement on the success in classification rates when these geometric features are considered

    Reinforcement Learning with Frontier-Based Exploration via Autonomous Environment

    Full text link
    Active Simultaneous Localisation and Mapping (SLAM) is a critical problem in autonomous robotics, enabling robots to navigate to new regions while building an accurate model of their surroundings. Visual SLAM is a popular technique that uses virtual elements to enhance the experience. However, existing frontier-based exploration strategies can lead to a non-optimal path in scenarios where there are multiple frontiers with similar distance. This issue can impact the efficiency and accuracy of Visual SLAM, which is crucial for a wide range of robotic applications, such as search and rescue, exploration, and mapping. To address this issue, this research combines both an existing Visual-Graph SLAM known as ExploreORB with reinforcement learning. The proposed algorithm allows the robot to learn and optimize exploration routes through a reward-based system to create an accurate map of the environment with proper frontier selection. Frontier-based exploration is used to detect unexplored areas, while reinforcement learning optimizes the robot's movement by assigning rewards for optimal frontier points. Graph SLAM is then used to integrate the robot's sensory data and build an accurate map of the environment. The proposed algorithm aims to improve the efficiency and accuracy of ExploreORB by optimizing the exploration process of frontiers to build a more accurate map. To evaluate the effectiveness of the proposed approach, experiments will be conducted in various virtual environments using Gazebo, a robot simulation software. Results of these experiments will be compared with existing methods to demonstrate the potential of the proposed approach as an optimal solution for SLAM in autonomous robotics.Comment: 23 pages, Journa

    A hybrid approach to simultaneous localization and mapping in indoors environment

    Get PDF
    This thesis will present SLAM in the current literature to benefit from then it will present the investigation results for a hybrid approach used where different algorithms using laser, sonar, and camera sensors were tested and compared. The contribution of this thesis is the development of a hybrid approach for SLAM that uses different sensors and where different factors are taken into consideration such as dynamic objects, and the development of a scalable grid map model with new sensors models for real time update of the map.The thesis will show the success found, difficulties faced and limitations of the algorithms developed which were simulated and experimentally tested in an indoors environment

    Robot training using system identification

    Get PDF
    This paper focuses on developing a formal, theory-based design methodology to generate transparent robot control programs using mathematical functions. The research finds its theoretical roots in robot training and system identification techniques such as Armax (Auto-Regressive Moving Average models with eXogenous inputs) and Narmax (Non-linear Armax). These techniques produce linear and non-linear polynomial functions that model the relationship between a robot’s sensor perception and motor response. The main benefits of the proposed design methodology, compared to the traditional robot programming techniques are: (i) It is a fast and efficient way of generating robot control code, (ii) The generated robot control programs are transparent mathematical functions that can be used to form hypotheses and theoretical analyses of robot behaviour, and (iii) It requires very little explicit knowledge of robot programming where end-users/programmers who do not have any specialised robot programming skills can nevertheless generate task-achieving sensor-motor couplings. The nature of this research is concerned with obtaining sensor-motor couplings, be it through human demonstration via the robot, direct human demonstration, or other means. The viability of our methodology has been demonstrated by teaching various mobile robots different sensor-motor tasks such as wall following, corridor passing, door traversal and route learning

    Automated steering design using Neural Network

    Get PDF
    If you don't move forward-you begin to move backward. Technological advancement today has brought us to a frontier where the human has become the basic constraint in our ascent towards safer and faster transportation. Human error is mostly responsible for many road traffic accidents which every year take the lives of lots of people and injure many more. Driving protection is thus a major concern leading to research in autonomous driving systems. Automatic motion planning and navigation is the primary task of an automated guided vehicle or mobile robots. All such navigation systems consist of a data collection system, a decision making system and a hardware control system. In this research our artificial intelligence system is based on neural network model for navigation of an AGV in unpredictable and imprecise environment. A five layered with gradient descent momentum back-propagation system which uses heading angle and obstacle distances as input. The networks are trained by real data obtained from vehicle tracking live test runs. Considering the high amount of risk of testing the vehicle in real space-time conditions, it would initially be tested in simulated environment with the use of MATLAB®. The hardware control for an AGV should be robust and precise. An Aerial and a Grounded prototype were developed to test our neural network model in real time situation

    Recognizing targets from infrared intensity scan patterns using artificial neural networks

    Get PDF
    This study investigates the use of simple, low-cost infrared sensors for the recognition of geometry and surface type of commonly encountered features or targets in indoor environments, such as planes, corners, and edges. The intensity measurements obtained from such sensors are highly dependent on the location, geometry, and surface properties of the reflecting target in a way that cannot be represented by a simple analytical relationship, therefore complicating the localization and recognition process. We employ artificial neural networks to determine the geometry and the surface type of targets and provide experimental verification with three different geometries and three different surface types. The networks are trained with the Levenberg-Marquardt algorithm and pruned with the optimal brain surgeon technique. The geometry and the surface type of targets can be correctly classified with rates of 99 and 78.4%, respectively. An average correct classification rate of 78% is achieved when both geometry and surface type are differentiated. This indicates that the geometrical properties of the targets are more distinctive than their surface properties, and surface determination is the limiting factor in recognizing the patterns. The results demonstrate that processing the data from simple infrared sensors through suitable techniques can help us exploit their full potential and extend their usage beyond well-known applications. © 2009 Society of Photo-Optical Instrumentation Engineers
    corecore