284 research outputs found

    System of Terrain Analysis, Energy Estimation and Path Planning for Planetary Exploration by Robot Teams

    Get PDF
    NASA’s long term plans involve a return to manned moon missions, and eventually sending humans to mars. The focus of this project is the use of autonomous mobile robotics to enhance these endeavors. This research details the creation of a system of terrain classification, energy of traversal estimation and low cost path planning for teams of inexpensive and potentially expendable robots. The first stage of this project was the creation of a model which estimates the energy requirements of the traversal of varying terrain types for a six wheel rocker-bogie rover. The wheel/soil interaction model uses Shibly’s modified Bekker equations and incorporates a new simplified rocker-bogie model for estimating wheel loads. In all but a single trial the relative energy requirements for each soil type were correctly predicted by the model. A path planner for complete coverage intended to minimize energy consumption was designed and tested. It accepts as input terrain maps detailing the energy consumption required to move to each adjacent location. Exploration is performed via a cost function which determines the robot’s next move. This system was successfully tested for multiple robots by means of a shared exploration map. At peak efficiency, the energy consumed by our path planner was only 56% that used by the best case back and forth coverage pattern. After performing a sensitivity analysis of Shibly’s equations to determine which soil parameters most affected energy consumption, a neural network terrain classifier was designed and tested. The terrain classifier defines all traversable terrain as one of three soil types and then assigns an assumed set of soil parameters. The classifier performed well over all, but had some difficulty distinguishing large rocks from sand. This work presents a system which successfully classifies terrain imagery into one of three soil types, assesses the energy requirements of terrain traversal for these soil types and plans efficient paths of complete coverage for the imaged area. While there are further efforts that can be made in all areas, the work achieves its stated goals

    Fully Proprioceptive Slip-Velocity-Aware State Estimation for Mobile Robots via Invariant Kalman Filtering and Disturbance Observer

    Full text link
    This paper develops a novel slip estimator using the invariant observer design theory and Disturbance Observer (DOB). The proposed state estimator for mobile robots is fully proprioceptive and combines data from an inertial measurement unit and body velocity within a Right Invariant Extended Kalman Filter (RI-EKF). By embedding the slip velocity into SE3(3)\mathrm{SE}_3(3) Lie group, the developed DOB-based RI-EKF provides real-time accurate velocity and slip velocity estimates on different terrains. Experimental results using a Husky wheeled robot confirm the mathematical derivations and show better performance than a standard RI-EKF baseline. Open source software is available for download and reproducing the presented results.Comment: github repository at https://github.com/UMich-CURLY/slip_detection_DOB. arXiv admin note: text overlap with arXiv:1805.10410 by other author

    Study of Mobile Robot Operations Related to Lunar Exploration

    Get PDF
    Mobile robots extend the reach of exploration in environments unsuitable, or unreachable, by humans. Far-reaching environments, such as the south lunar pole, exhibit lighting conditions that are challenging for optical imagery required for mobile robot navigation. Terrain conditions also impact the operation of mobile robots; distinguishing terrain types prior to physical contact can improve hazard avoidance. This thesis presents the conclusions of a trade-off that uses the results from two studies related to operating mobile robots at the lunar south pole. The lunar south pole presents engineering design challenges for both tele-operation and lidar-based autonomous navigation in the context of a near-term, low-cost, short-duration lunar prospecting mission. The conclusion is that direct-drive tele-operation may result in improved science data return. The first study is on demonstrating lidar reflectance intensity, and near-infrared spectroscopy, can improve terrain classification over optical imagery alone. Two classification techniques, Naive Bayes and multi-class SVM, were compared for classification errors. Eight terrain types, including aggregate, loose sand and compacted sand, are classified using wavelet-transformed optical images, and statistical values of lidar reflectance intensity. The addition of lidar reflectance intensity was shown to reduce classification errors for both classifiers. Four types of aggregate material are classified using statistical values of spectral reflectance. The addition of spectral reflectance was shown to reduce classification errors for both classifiers. The second study is on human performance in tele-operating a mobile robot over time-delay and in lighting conditions analogous to the south lunar pole. Round-trip time delay between operator and mobile robot leads to an increase in time to turn the mobile robot around obstacles or corners as operators tend to implement a `wait and see\u27 approach. A study on completion time for a cornering task through varying corridor widths shows that time-delayed performance fits a previously established cornering law, and that varying lighting conditions did not adversely affect human performance. The results of the cornering law are interpreted to quantify the additional time required to negotiate a corner under differing conditions, and this increase in time can be interpreted to be predictive when operating a mobile robot through a driving circuit

    Planetary Rover Inertial Navigation Applications: Pseudo Measurements and Wheel Terrain Interactions

    Get PDF
    Accurate localization is a critical component of any robotic system. During planetary missions, these systems are often limited by energy sources and slow spacecraft computers. Using proprioceptive localization (e.g., using an inertial measurement unit and wheel encoders) without external aiding is insufficient for accurate localization. This is mainly due to the integrated and unbounded errors of the inertial navigation solutions and the drifted position information from wheel encoders caused by wheel slippage. For this reason, planetary rovers often utilize exteroceptive (e.g., vision-based) sensors. On the one hand, localization with proprioceptive sensors is straightforward, computationally efficient, and continuous. On the other hand, using exteroceptive sensors for localization slows rover driving speed, reduces rover traversal rate, and these sensors are sensitive to the terrain features. Given the advantages and disadvantages of both methods, this thesis focuses on two objectives. First, improving the proprioceptive localization performance without significant changes to the rover operations. Second, enabling adaptive traversability rate based on the wheel-terrain interactions while keeping the localization reliable. To achieve the first objective, we utilized the zero-velocity, zero-angular rate updates, and non-holonomicity of a rover to improve rover localization performance even with the limited available sensor usage in a computationally efficient way. Pseudo-measurements generated from proprioceptive sensors when the rover is stationary conditions and the non-holonomic constraints while traversing can be utilized to improve the localization performance without any significant changes to the rover operations. Through this work, it is observed that a substantial improvement in localization performance, without the aid of additional exteroceptive sensor information. To achieve the second objective, the relationship between the estimation of localization uncertainty and wheel-terrain interactions through slip-ratio was investigated. This relationship was exposed with a Gaussian process with time series implementation by using the slippage estimation while the rover is moving. Then, it is predicted when to change from moving to stationary conditions by mapping the predicted slippage into localization uncertainty prediction. Instead of a periodic stopping framework, the method introduced in this work is a slip-aware localization method that enables the rover to stop more frequently in high-slip terrains whereas stops rover less frequently for low-slip terrains while keeping the proprioceptive localization reliable

    Terrain-Dependent Slip Risk Prediction for Planetary Exploration Rovers

    Get PDF
    Wheel slip prediction on rough terrain is crucial for secure, long-term operations of planetary exploration rovers. Although rough, unstructured terrain hampers mobility, prediction by modeling wheel–terrain interactions remains difficult owing to unclear terrain conditions and complexities of terramechanics models. This study proposes a vision-based approach with machine learning for predicting wheel slip risk by estimating the slope from 3D information and classifying terrain types from image information. It considers the slope estimation accuracy for risk prediction under sharp increases in wheel slip due to inclined ground. Experimental results obtained with a rover testbed on several terrain types validate this method

    Visual Prediction of Rover Slip: Learning Algorithms and Field Experiments

    Get PDF
    Perception of the surrounding environment is an essential tool for intelligent navigation in any autonomous vehicle. In the context of Mars exploration, there is a strong motivation to enhance the perception of the rovers beyond geometry-based obstacle avoidance, so as to be able to predict potential interactions with the terrain. In this thesis we propose to remotely predict the amount of slip, which reflects the mobility of the vehicle on future terrain. The method is based on learning from experience and uses visual information from stereo imagery as input. We test the algorithm on several robot platforms and in different terrains. We also demonstrate its usefulness in an integrated system, onboard a Mars prototype rover in the JPL Mars Yard. Another desirable capability for an autonomous robot is to be able to learn about its interactions with the environment in a fully automatic fashion. We propose an algorithm which uses the robot's sensors as supervision for vision-based learning of different terrain types. This algorithm can work with noisy and ambiguous signals provided from onboard sensors. To be able to cope with rich, high-dimensional visual representations we propose a novel, nonlinear dimensionality reduction technique which exploits automatic supervision. The method is the first to consider supervised nonlinear dimensionality reduction in a probabilistic framework using supervision which can be noisy or ambiguous. Finally, we consider the problem of learning to recognize different terrains, which addresses the time constraints of an onboard autonomous system. We propose a method which automatically learns a variable-length feature representation depending on the complexity of the classification task. The proposed approach achieves a good trade-off between decrease in computational time and recognition performance.</p

    Rut detection and following for autonomous ground vehicles

    Full text link
    corecore