143 research outputs found

    3D Maize Plant Reconstruction Based on Georeferenced Overlapping LiDAR Point Clouds

    Get PDF
    3D crop reconstruction with a high temporal resolution and by the use of non-destructive measuring technologies can support the automation of plant phenotyping processes. Thereby, the availability of such 3D data can give valuable information about the plant development and the interaction of the plant genotype with the environment. This article presents a new methodology for georeferenced 3D reconstruction of maize plant structure. For this purpose a total station, an IMU, and several 2D LiDARs with different orientations were mounted on an autonomous vehicle. By the multistep methodology presented, based on the application of the ICP algorithm for point cloud fusion, it was possible to perform the georeferenced point clouds overlapping. The overlapping point cloud algorithm showed that the aerial points (corresponding mainly to plant parts) were reduced to 1.5%–9% of the total registered data. The remaining were redundant or ground points. Through the inclusion of different LiDAR point of views of the scene, a more realistic representation of the surrounding is obtained by the incorporation of new useful information but also of noise. The use of georeferenced 3D maize plant reconstruction at different growth stages, combined with the total station accuracy could be highly useful when performing precision agriculture at the crop plant level

    UAV Oblique Imagery with an Adaptive Micro-Terrain Model for Estimation of Leaf Area Index and Height of Maize Canopy from 3D Point Clouds

    Get PDF
    Leaf area index (LAI) and height are two critical measures of maize crops that are used in ecophysiological and morphological studies for growth evaluation, health assessment, and yield prediction. However, mapping spatial and temporal variability of LAI in fields using handheld tools and traditional techniques is a tedious and costly pointwise operation that provides information only within limited areas. The objective of this study was to evaluate the reliability of mapping LAI and height of maize canopy from 3D point clouds generated from UAV oblique imagery with the adaptive micro-terrain model. The experiment was carried out in a field planted with three cultivars having different canopy shapes and four replicates covering a total area of 48 Ă— 36 m. RGB images in nadir and oblique view were acquired from the maize field at six different time slots during the growing season. Images were processed by Agisoft Metashape to generate 3D point clouds using the structure from motion method and were later processed by MATLAB to obtain clean canopy structure, including height and density. The LAI was estimated by a multivariate linear regression model using crop canopy descriptors derived from the 3D point cloud, which account for height and leaf density distribution along the canopy height. A simulation analysis based on the Sine function effectively demonstrated the micro-terrain model from point clouds. For the ground truth data, a randomized block design with 24 sample areas was used to manually measure LAI, height, N-pen data, and yield during the growing season. It was found that canopy height data from the 3D point clouds has a relatively strong correlation (R2 = 0.89, 0.86, 0.78) with the manual measurement for three cultivars with CH90 . The proposed methodology allows a cost-effective high-resolution mapping of in-field LAI index extraction through UAV 3D data to be used as an alternative to the conventional LAI assessments even in inaccessible regions

    Using Assembled 2D LiDAR Data for Single Plant Detection

    Full text link
    A 2D laser scanner was mounted on the front of the small 4-wheel autonomous robot with differential steering, in an angle of 30 degrees pointing downwards. The machine was able to drive between maize rows and collect timestamped data simultaneously. The position of the vehicle was tracked by a highly precise total station. The data of the total station and the laser scanner was fused to generate a 3D point cloud. This 3D representation was used to search for single plant positions, what could later be used for additional applications like single plant treatment and precision weeding. First all points belonging to the ground plane were removed. Afterwards outliers were filtered. For separating the resulting points, a k-d tree clustering was used. Of each single point cloud cluster the 3D centroid was evaluated and assumed as the resulting plant position. This was done on three different growth stages of the plants. Results showed good detection rates up to 70.7 % with a root mean square error of 3.6 cm, precise enough to allow single plant treatment

    Crop plant reconstruction and feature extraction based on 3-D vision

    Get PDF
    3-D imaging is increasingly affordable and offers new possibilities for a more efficient agricul-tural practice with the use of highly advances technological devices. Some reasons contrib-uting to this possibility include the continuous increase in computer processing power, the de-crease in cost and size of electronics, the increase in solid state illumination efficiency and the need for greater knowledge and care of the individual crops. The implementation of 3-D im-aging systems in agriculture is impeded by the economic justification of using expensive de-vices for producing relative low-cost seasonal products. However, this may no longer be true since low-cost 3-D sensors, such as the one used in this work, with advance technical capabili-ties are already available. The aim of this cumulative dissertation was to develop new methodologies to reconstruct the 3-D shape of agricultural environment in order to recognized and quantitatively describe struc-tures, in this case: maize plants, for agricultural applications such as plant breeding and preci-sion farming. To fulfil this aim a comprehensive review of the 3-D imaging systems in agricul-tural applications was done to select a sensor that was affordable and has not been fully inves-tigated in agricultural environments. A low-cost TOF sensor was selected to obtain 3-D data of maize plants and a new adaptive methodology was proposed for point cloud rigid registra-tion and stitching. The resulting maize 3-D point clouds were highly dense and generated in a cost-effective manner. The validation of the methodology showed that the plants were recon-structed with high accuracies and the qualitative analysis showed the visual variability of the plants depending on the 3-D perspective view. The generated point cloud was used to obtain information about the plant parameters (stem position and plant height) in order to quantita-tively describe the plant. The resulting plant stem positions were estimated with an average mean error and standard deviation of 27 mm and 14 mm, respectively. Additionally, meaning-ful information about the plant height profile was also provided, with an average overall mean error of 8.7 mm. Since the maize plants considered in this research were highly heterogeneous in height, some of them had folded leaves and were planted with standard deviations that emulate the real performance of a seeder; it can be said that the experimental maize setup was a difficult scenario. Therefore, a better performance, for both, plant stem position and height estimation could be expected for a maize field in better conditions. Finally, having a 3-D re-construction of the maize plants using a cost-effective sensor, mounted on a small electric-motor-driven robotic platform, means that the cost (either economic, energetic or time) of gen-erating every point in the point cloud is greatly reduced compared with previous researches.Die 3D-Bilderfassung ist zunehmend kostengünstiger geworden und bietet neue Möglichkeiten für eine effizientere landwirtschaftliche Praxis durch den Einsatz hochentwickelter technologischer Geräte. Einige Gründe, die diese ermöglichen, ist das kontinuierliche Wachstum der Computerrechenleistung, die Kostenreduktion und Miniaturisierung der Elektronik, die erhöhte Beleuchtungseffizienz und die Notwendigkeit einer besseren Kenntnis und Pflege der einzelnen Pflanzen. Die Implementierung von 3-D-Sensoren in der Landwirtschaft wird durch die wirtschaftliche Rechtfertigung der Verwendung teurer Geräte zur Herstellung von kostengünstigen Saisonprodukten verhindert. Dies ist jedoch nicht mehr länger der Fall, da kostengünstige 3-D-Sensoren, bereits verfügbar sind. Wie derjenige dier in dieser Arbeit verwendet wurde. Das Ziel dieser kumulativen Dissertation war, neue Methoden für die Visualisierung die 3-D-Form der landwirtschaftlichen Umgebung zu entwickeln, um Strukturen quantitativ zu beschreiben: in diesem Fall Maispflanzen für landwirtschaftliche Anwendungen wie Pflanzenzüchtung und Precision Farming zu erkennen. Damit dieses Ziel erreicht wird, wurde eine umfassende Überprüfung der 3D-Bildgebungssysteme in landwirtschaftlichen Anwendungen durchgeführt, um einen Sensor auszuwählen, der erschwinglich und in landwirtschaftlichen Umgebungen noch nicht ausgiebig getestet wurde. Ein kostengünstiger TOF-Sensor wurde ausgewählt, um 3-D-Daten von Maispflanzen zu erhalten und eine neue adaptive Methodik wurde für die Ausrichtung von Punktwolken vorgeschlagen. Die resultierenden Mais-3-D-Punktwolken hatten eine hohe Punktedichte und waren in einer kosteneffektiven Weise erzeugt worden. Die Validierung der Methodik zeigte, dass die Pflanzen mit hoher Genauigkeit rekonstruiert wurden und die qualitative Analyse die visuelle Variabilität der Pflanzen in Abhängigkeit der 3-D-Perspektive zeigte. Die erzeugte Punktwolke wurde verwendet, um Informationen über die Pflanzenparameter (Stammposition und Pflanzenhöhe) zu erhalten, die die Pflanze quantitativ beschreibt. Die resultierenden Pflanzenstammpositionen wurden mit einem durchschnittlichen mittleren Fehler und einer Standardabweichung von 27 mm bzw. 14 mm berechnet. Zusätzlich wurden aussagekräftige Informationen zum Pflanzenhöhenprofil mit einem durchschnittlichen Gesamtfehler von 8,7 mm bereitgestellt. Da die untersuchten Maispflanzen in der Höhe sehr heterogen waren, hatten einige von ihnen gefaltete Blätter und wurden mit Standardabweichungen gepflanzt, die die tatsächliche Genauigkeit einer Sämaschine nachahmen. Man kann sagen, dass der experimentelle Versuch ein schwieriges Szenario war. Daher könnte für ein Maisfeld unter besseren Bedingungen eine besseres Resultat sowohl für die Pflanzenstammposition als auch für die Höhenschätzung erwartet werden. Schließlich bedeutet eine 3D-Rekonstruktion der Maispflanzen mit einem kostengünstigen Sensor, der auf einer kleinen elektrischen, motorbetriebenen Roboterplattform montiert ist, dass die Kosten (entweder wirtschaftlich, energetisch oder zeitlich) für die Erzeugung jedes Punktes in den Punktwolken im Vergleich zu früheren Untersuchungen stark reduziert werden

    Methods and Applications of 3D Ground Crop Analysis Using LiDAR Technology: A Survey

    Get PDF
    Light Detection and Ranging (LiDAR) technology is positioning itself as one of the most effective non-destructive methods to collect accurate information on ground crop fields, as the analysis of the three-dimensional models that can be generated with it allows for quickly measuring several key parameters (such as yield estimations, aboveground biomass, vegetation indexes estimation, perform plant phenotyping, and automatic control of agriculture robots or machinery, among others). In this survey, we systematically analyze 53 research papers published between 2005 and 2022 that involve significant use of the LiDAR technology applied to the three-dimensional analysis of ground crops. Different dimensions are identified for classifying the surveyed papers (including application areas, crop species under study, LiDAR scanner technologies, mounting platform technologies, and the use of additional instrumentation and software tools). From our survey, we draw relevant conclusions about the use of LiDAR technologies, such as identifying a hierarchy of different scanning platforms and their frequency of use as well as establishing the trade-off between the economic costs of deploying LiDAR and the agronomically relevant information that effectively can be acquired. We also conclude that none of the approaches under analysis tackles the problem associated with working with multiple species with the same setup and configuration, which shows the need for instrument calibration and algorithmic fine tuning for an effective application of this technology.Fil: Micheletto, Matías Javier. Consejo Nacional de Investigaciones Cientificas y Tecnicas. Centro de Investigaciones y Transferencia Golfo San Jorge. Centro de Investigaciones y Transferencia Golfo San Jorge: Sede Caleta Olivia - Santa Cruz | Universidad Nacional de la Patagonia Austral. Centro de Investigaciones y Transferencia Golfo San Jorge. Centro de Investigaciones y Transferencia Golfo San Jorge: Sede Caleta Olivia - Santa Cruz | Universidad Nacional de la Patagonia "san Juan Bosco". Centro de Investigaciones y Transferencia Golfo San Jorge. Centro de Investigaciones y Transferencia Golfo San Jorge: Sede Caleta Olivia - Santa Cruz; ArgentinaFil: Chesñevar, Carlos Iván. Universidad Nacional del Sur. Departamento de Ciencias e Ingeniería de la Computación; Argentina. Consejo Nacional de Investigaciones Científicas y Técnicas. Centro Científico Tecnológico Conicet - Bahía Blanca. Instituto de Ciencias e Ingeniería de la Computación. Universidad Nacional del Sur. Departamento de Ciencias e Ingeniería de la Computación. Instituto de Ciencias e Ingeniería de la Computación; ArgentinaFil: Santos, Rodrigo Martin. Consejo Nacional de Investigaciones Científicas y Técnicas. Centro Científico Tecnológico Conicet - Bahía Blanca. Instituto de Ciencias e Ingeniería de la Computación. Universidad Nacional del Sur. Departamento de Ciencias e Ingeniería de la Computación. Instituto de Ciencias e Ingeniería de la Computación; Argentina. Universidad Nacional del Sur. Departamento de Ingeniería Eléctrica y de Computadoras; Argentin

    Estimating Canopy Parameters Based on the Stem Position in Apple Trees Using a 2D LiDAR

    Get PDF
    Data of canopy morphology are crucial for cultivation tasks within orchards. In this study, a 2D light detection and range (LiDAR) laser scanner system was mounted on a tractor, tested on a box with known dimensions (1.81 m Ă— 0.6 m Ă— 0.6 m), and applied in an apple orchard to obtain the 3D structural parameters of the trees (n = 224). The analysis of a metal box which considered the height of four sides resulted in a mean absolute error (MAE) of 8.18 mm with a bias (MBE) of 2.75 mm, representing a root mean square error (RMSE) of 1.63% due to gaps in the point cloud and increased incident angle with enhanced distance between laser aperture and the object. A methodology based on a bivariate point density histogram is proposed to estimate the stem position of each tree. The cylindrical boundary was projected around the estimated stem positions to segment each individual tree. Subsequently, height, stem diameter, and volume of the segmented tree point clouds were estimated and compared with manual measurements. The estimated stem position of each tree was defined using a real time kinematic global navigation satellite system, (RTK-GNSS) resulting in an MAE and MBE of 33.7 mm and 36.5 mm, respectively. The coefficient of determination (R2) considering manual measurements and estimated data from the segmented point clouds appeared high with, respectively, R2 and RMSE of 0.87 and 5.71% for height, 0.88 and 2.23% for stem diameter, as well as 0.77 and 4.64% for canopy volume. Since a certain error for the height and volume measured manually can be assumed, the LiDAR approach provides an alternative to manual readings with the advantage of getting tree individual data of the entire orchard

    Acquiring plant features with optical sensing devices in an organic strip-cropping system

    Get PDF
    The SUREVEG project focuses on improvement of biodiversity and soil fertility in organic agriculture through strip-cropping systems. To counter the additional workforce a robotic tool is proposed. Within the project, a modular proof of concept (POC) version will be produced that will combine detection technologies with actuation on a single-plant level in the form of a robotic arm. This article focuses on the detection of crop characteristics through point clouds obtained with two lidars. Segregation in soil and plants was successfully achieved without the use of additional data from other sensor types, by calculating weighted sums, resulting in a dynamically obtained threshold criterion. This method was able to extract the vegetation from the point cloud in strips with varying vegetation coverage and sizes. The resulting vegetation clouds were compared to drone imagery, to prove they perfectly matched all green areas in said image. By dividing the remaining clouds of overlapping plants by means of the nominal planting distance, the number of plants, their volumes, and thereby the expected yields per row could be determined.</p
    • …
    corecore