1,362 research outputs found

    Disease Detection of Solanaceous Crops Using Deep Learning for Robot Vision

    Get PDF
    Traditionally, the farmers manage the crops from the early growth stage until the mature harvest stage by manually identifying and monitoring plant diseases, nutrient deficiencies, controlled irrigation, and controlled fertilizers and pesticides. Even the farmers have difficulty detecting crop diseases using their naked eyes due to several similar crop diseases. Identifying the correct diseases is crucial since it can improve the quality and quantity of crop production. With the advent of Artificial Intelligence (AI) technology, all crop-managing tasks can be automated using a robot that mimics a farmer's ability. However, designing a robot with human capability, especially in detecting the crop's diseases in real-time, is another challenge to consider. Other research works are focusing on improving the mean average precision and the best result reported so far is 93% of mean Average Precision (mAP) by YOLOv5. This paper focuses on object detection of the Convolutional Neural Network (CNN) architecture-based to detect the disease of solanaceous crops for robot vision. This study's contribution involved reporting the developmental specifics and a suggested solution for issues that appear along with the conducted study. In addition, the output of this study is expected to become the algorithm of the robot's vision. This study uses images of four crops (tomato, potato, eggplant, and pepper), including 23 classes of healthy and diseased crops infected on the leaf and fruits. The dataset utilized combines the public dataset (PlantVillage) and self-collected samples. The total dataset of all 23 classes is 16580 images divided into three parts: training set, validation set, and testing set. The dataset used for training is 88% of the total dataset (15000 images), 8% of the dataset performed a validation process (1400 images), and the rest of the 4% dataset is for the test process (699 images). The performances of YOLOv5 were more robust in terms of 94.2% mAP, and the speed was slightly faster than Scaled-YOLOv4. This object detection-based approach has proven to be a promising solution in efficiently detecting crop disease in real-time

    Complexer-YOLO: Real-Time 3D Object Detection and Tracking on Semantic Point Clouds

    Full text link
    Accurate detection of 3D objects is a fundamental problem in computer vision and has an enormous impact on autonomous cars, augmented/virtual reality and many applications in robotics. In this work we present a novel fusion of neural network based state-of-the-art 3D detector and visual semantic segmentation in the context of autonomous driving. Additionally, we introduce Scale-Rotation-Translation score (SRTs), a fast and highly parameterizable evaluation metric for comparison of object detections, which speeds up our inference time up to 20\% and halves training time. On top, we apply state-of-the-art online multi target feature tracking on the object measurements to further increase accuracy and robustness utilizing temporal information. Our experiments on KITTI show that we achieve same results as state-of-the-art in all related categories, while maintaining the performance and accuracy trade-off and still run in real-time. Furthermore, our model is the first one that fuses visual semantic with 3D object detection

    An End-to-End Approach for Recognition of Modern and Historical Handwritten Numeral Strings

    Full text link
    An end-to-end solution for handwritten numeral string recognition is proposed, in which the numeral string is considered as composed of objects automatically detected and recognized by a YoLo-based model. The main contribution of this paper is to avoid heuristic-based methods for string preprocessing and segmentation, the need for task-oriented classifiers, and also the use of specific constraints related to the string length. A robust experimental protocol based on several numeral string datasets, including one composed of historical documents, has shown that the proposed method is a feasible end-to-end solution for numeral string recognition. Besides, it reduces the complexity of the string recognition task considerably since it drops out classical steps, in special preprocessing, segmentation, and a set of classifiers devoted to strings with a specific length

    Automatic target recognition with convolutional neural networks.

    Get PDF
    Automatic Target Recognition (ATR) characterizes the ability for an algorithm or device to identify targets or other objects based on data obtained from sensors, being commonly thermal. ATR is an important technology for both civilian and military computer vision applications. However, the current level of performance that is available is largely deficient compared to the requirements. This is mainly due to the difficulty of acquiring targets in realistic environments, and also to limitations of the distribution of classified data to the academic community for research purposes. This thesis proposes to solve the ATR task using Convolutional Neural Networks (CNN). We present three learning approaches using WideResNet-28-2\cite{wrn} as a backbone CNN. The first method uses random initialization of the network weights. The second method explores transfer learning. Finally, the third approach relies on spatial transformer networks \cite{stn} to enhance the geometric invariance of the model. To validate, analyze and compare our three proposed models, we use a large-scale real benchmark dataset that includes civilian and military vehicles. These targets are captured at different viewing angles, different resolutions, and different times of the day. We evaluate the effectiveness of our methods by studying their robustness to realistic case scenarios where no ground truth data is available and targets are automatically detected. We show that the method that uses spatial transformer networks achieves the best results and demonstrates the most robustness to various perturbations that can be encountered in real applications
    • …
    corecore