194 research outputs found

    Chili plant classification using transfer learning models through object detection

    Get PDF
    This study presents an application of using a Convolutional Neural Network (CNN) based detector to detect chili and its leaves in the chili plant image. Detecting chili on its plant is essential for the development of robotic vision and monitoring. Thus, helps us supervise the plant growth, furthermore, analyses their productivity and quality. This paper aims to develop a system that can monitor and identify bird’s eye chili plants by implementing machine learning. First, the development of methodology for efficient detection of bird’s eye chili and its leaf was made. A dataset of a total of 1866 images after augmentation of bird’s eye chili and its leaf was used in this experiment. YOLO Darknet was implemented to train the dataset. After a series of experiments were conducted, the model is compared with other transfer learning models like YOLO Tiny, Faster R-CNN, and EfficientDet. The classification performance of these transfer learning models has been calculated and compared with each other. The experimental result shows that the Yolov4 Darknet model achieves mAP of 75.69%, followed by EfficientDet at 71.85% for augmented dataset

    Deep learning for medium-scale agricultural crop detection through aerial view images

    Get PDF
    This research project focuses on utilizing two state-of-the-art YOLOv4-based deep learning models, for large-scale agricultural crop detection using Unmanned Aerial Vehicles (UAVs). The objective is to develop an accurate and efficient crop detection system capable of identifying chili crops, eggplant crops, and empty polybags in agricultural fields. Crops detection is important for the development of a robotic vision in maximize the productivity and efficiency in agriculture associate with the development of concept Industry 4.0. This study seek to explore the comparison between YOLOv4 and YOLOv4 tiny model in term of mean average precision (mAP), precision, recall, F1-score, detection time and memory consumption. A custom dataset with 300 images was collected and annotated into total bounding boxes of 23335 with 6969 chili tree, 15402 eggplant tree and 964 empty polybag. The dataset was separated into train, validation and test set with the ratio of 70:20:10. The dataset was trained into YOLOv4 and YOLOv4 tiny with 2000 iterations. The result has shown that the YOLOv4 has the higher mean AP of 91.49% with 244.2mb memory storage consumption while YOLOv4 tiny achieve lower mean AP of 71.83% with 22.4mb. In summary, this research has significated the implementation of deep learning models to perform large-scale agricultural crop detection and can be further develop into automation industrial 4.0 of local agricultural sector

    Features extraction of capsicum frutescens (C.F) NDVI values using image processing

    Get PDF
    There is yet an application for monitoring plant condition using the Normalized Difference Vegetation Index (NDVI) method for Capsicum Frutescens (C.F) or chili. This study was carried out in three phases, where the first and second phases are to create NDVI images and recognize and extract features from NDVI images. The last stage is to assess the efficiency of Neural Network (N.N.), Naïve Bayes (N.B.), and Logistic Regression (L.R.) models on the classification of chili plant health. The images of the chili plant will be captured using two types of cameras, which can be differentiated by whether or not they have an infrared filter. The images were collected to create datasets, and the NDVI images' features were extracted. The 120 NDVI images of the chili plant were divided into training and test datasets, with 70.0% training and 30.0% test. The extracted data was used to test the classification accuracy of classifiers on datasets. Finally, the N.N. model was found to have the highest classification accuracy, with 96.4 % on the training dataset and 88.9 % on the test dataset. The state of the chili plant can be predicted based on feature extraction from NDVI images by the end of the study

    Comparison of different deep learning object detection algorithms on fruit drying characterization

    Get PDF
    Object detection is an essential task in the field of computer vision and a prominent area of research. In the past, the categorization of raw and dry Tamanu fruits was dependent on human perception. Nevertheless, due to the progress in object detection, this task can currently be computerized. This study employs three deep learning object detection models: You Only Look Once v5m (YOLOv5m), Single Shot Detector (SSD) MobileNet and EfficientDet. The models were trained using images of Tamanu fruits in their raw and dry state, which were directly collected from the dryer device. Following the completion of training, the models underwent evaluation to identify the one with the highest level of accuracy. YOLOv5m demonstrated superior performance compared to SSD MobileNet and EfficientDet, achieving a mean average precision (mAP) of 0.99589. SSD MobileNet demonstrated exceptional performance in real-time object detection, accurately detecting the majority of objects with a high level of confidence. This study showcases the efficacy of employing deep learning object detection models to automate the classification of Tamanu fruit

    The diagnosis of diabetic retinopathy by means of transfer learning and fine-tuned dense layer pipeline

    Get PDF
    Diabetes is a global disease that occurs when the body is disabled pancreas to secrete insulin to convert the sugar to power in the blood. As a result, some tiny blood vessels on the part of the body, such as the eyes, are affected by high sugar and cause blocking blood flow in the vessels, which is called diabetic retinopathy. This disease may lead to permanent blindness due to the growth of new vessels in the back of the retina causing it to detach from the eyes. In 2016, 387 million people were diagnosed with Diabetic retinopathy, and the number is growing yearly, and the old detection approach becomes worse. Therefore, the purpose of this paper is to computerize the old method of detecting different classes of DR from 0-4 according to severity by given fundus images. The method is to construct a fine-tuned deep learning model based on transfer learning with dense layers. The used models here are InceptionV3, VGG16, and ResNet50 with a sharpening filter. Subsequently, InceptionV3 has achieved 94% as the highest accuracy among other models

    The identification of significant features towards travel mode choice and its prediction via optimised random forest classifier : An evaluation for active commuting behavior

    Get PDF
    Physical activity is the foundation to staying healthy, but sedentary activities have become not uncommon that ought to be mitigated immediately. The study aims to highlight the role of a transport system that encourages physical activity among users by applying an active door-to-door transport system. Users’ mode choice is studied to understand their preferences for active commuting. The use of machine learning has since been ubiquitous in a myriad of fields, including transportation studies and hence is also investigated towards its efficacy in predicting travel mode choice. Methodology: The application of the Random Forest (RF) model to identify travel mode choice is explored using the Revealed/Stated Preferences (RP/SP) Survey data in Kuantan City during weekdays. A total of 386 respondents were involved in this survey. The efficacy of the tuned RF models towards predicting the travel mode choice is evaluated via the Classification Accuracy (CA) performance indicator. In addition, a Feature Importance study is also carried out in order to identify significant factors that contribute towards travel mode choice. Results: The results from the present investigation demonstrated that the default RF model has acceptable predictability for both training and test dataset of users’ mode choice, with a CA of 70.2% and 69.3%, respectively. Upon identifying the significant features and further refining the hyperparameters of the RF model heuristically, it was shown that with 145 trees, the CA improved to up to 71.6% and 70.1% for both the training and test dataset, respectively. Through the feature selection technique, the most significant features that affect users mode choice are total travel time (TT), waiting time at a public transport stop (WT), region, walking distance from the last stop to destination (WD2), and walking distance from home to the nearest bus stop (WD1). Conclusions: The study has illustrated the efficacy of the optimised RF in predicting travel mode choice as well as identified the significant factors for the selection. The findings of the present study provide significant insight for policymakers to improve the performance of the public transportation system so that the users will benefit in terms of health and well-being from active commuting

    The classification of heart murmurs: The identification of significant time domain features

    Get PDF
    Phonocardiogram (PCG) is a type of acoustic signal collected from the heartbeat sound. PCG signals collected in the form of wave files and collected type of heart sound with a specific period. The difficulty of the binomial class in supervised machine learning is the steps-by-steps workflow. The consideration and decision make for every part are importantly stated so that misclassification will not occur. For the heart murmurs classification, data extraction has highly cared for it as we might have fault data consisting of outside signals. Before classifying murmurs in binomial, it will involve multiple features for selection that can have a better classification of the heart murmurs. Nevertheless, since classification performance is vital to conclude the results, models are needed to compare the research's output. The main objective of this study is to classify the signal of the murmur via time-domain based EEG signals. In this study, significant time-domain features were identified to determine the best features by using different feature selection methods. It continues with the classification with different models to compete for the highest accuracy as the performance for murmur classification. A set of Michigan Heart Sound and Murmur database (MHSDB) was provided by the University of Michigan Health System with chosen signals listening with the bell of the stethoscope at the apex area by left decubitus posture of the subjects. The PCG signals are pre-processed by segmentation of three seconds, downsampling eight thousand Hz and normalized to -1, +1. Features are extracted into ten features: Root Mean Square, Variance, Standard Deviation, Maximum, Minimum, Median, Skewness, Shape Factor, Kurtosis, and Mean. Two cross-validation methods applied, such as data splitting and k-fold cross-validation, were used to measure this study's data. Chi-Square and ANOVA technique practice to identify the significant features to improve the classification performance. The classification learners are enrolled and compared by Ada Boost, Random Forest (RF) and Support Vector Machine (SVM). The datasets will separate into a ratio of 70:30 and 80:20 for training and testing data, respectively. The chi-Square selection method was the best features selection method and 80:20 data splitting with better performance than the 70:30 ratio. The best classification accuracy for the models significantly come by SVM with all the categories with 100% except 70:30 test on test data with 97.2% only

    Screw absence classification on aluminum plate via features based transfer learning models

    Get PDF
    Screw is one of the important elements in every industry. Present of screw play an important role in which it holds the product in its own position and prevent loosen or collision with the case which will cause the small components or compartment fall off from its original position and lead to product failure. With the rise of revolution 4.0 in the industry, it helps to reduce the labor cost and human error. The main purpose of this study is to create a robust classification model used for machine vision detection – absence and present of screw, which could be adapted into respective robotics application system. 6 degree of freedom UR robot, Universal Robot is used to collect the custom dataset in TT Vision Technologies Sdn Bhd. The collected dataset is then classified into two categories, named as absent and present. Pretrained dataset, ImageNet is used to ease the training process in this research. Transfer learning model is used to extract the features which used to feed into different machine learning models. Each machine learning models undergoes hyperparameters tunning to achieve best classification accuracy. Samling ratio of 60:20:20 is used to separate the data in training, validation and testing respectively before fed into different ml model

    Articulated robot arm

    Get PDF
    In medical rehabilitation programs, trajectory tracking is used to increase the repeatability of joint movement and the patient's recovery in the early phases of rehabilitation. In order to achieve that, the robotic arm has been implemented since it can provide a precise and move in almost perfect motion. This manuscript aim to develop and simulate a 2DOF robotic arn that will able to tracking the trajectory successfully. Hence, in order to achieved that a modeling, simulation, and control of a Two Degree of Freedom (2-DOF) Robot Arm is being discussed in this manuscript. First, the robot specifications, as well as Robot Kinematics forward and inverse kinematics of a 2-DOF robot arm, are provided. The dynamics of the 2-DOF robot arm were then formulated in order to obtain motion equations by using the Eular-Lagrange Equation. For the controller of the robot, a control design was created utilising a PID controller. All the data is recorded from the margin of error as well as the overshoot and peak settling time is being record via matlab. The data is differentiate by with with controller, with PI and PID, in which the error is less than 12.5 and 1.63 consecutively. The data that being gathered show that a controller best suited in this rehabilitation robo

    Sign language recognition using deep learning through LSTM and CNN

    Get PDF
    This study presents the application of using deep learning to detect, recognize and translate sign language. Understanding sign language is crucial for communication between the deaf and mute people and the general society. This helps sign language users to easily communicate with others, thus eliminating the differences between both parties. The objectives of this thesis are to extract features from the dataset for sign language recognition model and the formulation of deep learning models and the classification performance to carry out the sign language recognition. First, we develop methodology for an efficient recognition of sign language. Next is to develop multiple system using three different model which is LSTM, CNN and YOLOv5 and compare the real time test result to choose the best model with the highest accuracy. We used same datasets for all algorithms to determine the best algorithm. The YOLOv5 has achieved the highest accuracy of 97% followed by LSTM and CNN with 94% and 66.67%
    corecore