148,678 research outputs found

    Augmented MRI Images for Classification of Normal and Tumors Brain through Transfer Learning Techniques

    Get PDF
    A brain tumor is a severe malignant condition caused by uncontrolled and abnormal cell division. Recent advances in deep learning have aided the health business in Medical Imaging for the diagnosis of numerous disorders. The most frequent and widely used deep learning algorithm for visual learning and image recognition. This research seeks to multi-classification tumors in the brain from images attained by Magnetic Resonance Imaging (MRI) using deep learning models that have been pre-trained for transfer learning. As per the publicly available MRI brain tumor dataset, brain tumors identified as glioma, meningioma, and pituitary, are accounting for most brain tumors. To ensure the robustness of the suggested method, data acquisition, and preprocessing are performed in the first step followed by data augmentation. Finally, Transfer Learning algorithms including DenseNet, ResNetV2, and InceptionResNetv2 have been applied to find out the optimum algorithm based on various parameters including accuracy, precision, and recall, and are under the curve (AUC). The experimental outcomes show that the model’s validation accuracy is high for DenseNet (about 97%), while ResNetv2 and InceptionResNetv2 achieved 77% and 80% only

    A Novel Method for Fashion Clothing Image Classification Based on Deep Learning

    Get PDF
    Image recognition and classification is a significant research topic in computational vision and widely used computer technology. The methods often used in image classification and recognition tasks are based on deep learning, like Convolutional Neural Networks (CNNs), LeNet, and Long Short-Term Memory networks (LSTM). Unfortunately, the classification accuracy of these methods is unsatisfactory. In recent years, using large-scale deep learning networks to achieve image recognition and classification can improve classification accuracy, such as VGG16 and Residual Network (ResNet). However, due to the deep network hierarchy and complex parameter settings, these models take more time in the training phase, especially when the sample number is small, which can easily lead to overfitting. This paper suggested a deep learning-based image classification technique based on a CNN model and improved convolutional and pooling layers. Furthermore, the study adopted the approximate dynamic learning rate update algorithm in the model training to realize the learning rate’s self-adaptation, ensure the model’s rapid convergence, and shorten the training time. Using the proposed model, an experiment was conducted on the Fashion-MNIST dataset, taking 6,000 images as the training dataset and 1,000 images as the testing dataset. In actual experiments, the classification accuracy of the suggested method was 93 percent, 4.6 percent higher than that of the basic CNN model. Simultaneously, the study compared the influence of the batch size of model training on classification accuracy. Experimental outcomes showed this model is very generalized in fashion clothing image classification tasks

    Deep learning based vision inspection system for remanufacturing application

    Get PDF
    Deep Learning has emerged as a state-of-the-art learning technique across a wide range of applications, including image recognition, localisation, natural language processing, prediction and forecasting systems. With significant applicability, Deep Learning is continually seeking other new fronts of applications for these techniques. This research is the first to apply Deep Learning algorithm to inspection in remanufacturing. Inspection is a key process in remanufacturing, which is currently an expensive manual operation in the remanufacturing process that depends on human operator expertise, in most cases. This research further proposes an automation framework based on Deep Learning algorithm for automating this inspection process. The proposed technique offers the potential to eliminate human factors in inspection, save cost, increase throughput and improve precision. This paper presents a novel vision-based inspection system on Deep Convolution Neural Network (DCNN) for three types of defects, namely pitting, surface abrasion and cracks by distinguishing between these surface defected parts. The materials used for this feasibility study were 100cm x 150cm mild steel plate material, purchased locally, and captured using a web webcam USB camera of 0.3 megapixels. The performance of this preliminary study indicates that the DCNN can classify with up to 100% accuracy on validation data and above 96% accuracy on a live video feed, by using 80% of the sample dataset for training and the remaining 20% for testing. Therefore, in the remanufacturing parts inspection, the DCNN approach has high potential as a method that could surpass the current technologies, especially for accuracy and speed. This preliminary study demonstrates that Deep Learning techniques have the potential to revolutionise inspection in remanufacturing. This research offers valuable insight into these opportunities, serving as a starting point for future applications of Deep Learning algorithms to remanufacturing

    SMART STRUCTURAL CONDITION ASSESSMENT METHODS FOR CIVIL INFRASTRUCTURES USING DEEP LEARNING ALGORITHM

    Get PDF
    Smart Structural Health Monitoring (SHM) technique capable of automated and accurate structural health condition assessment is appealing since civil infrastructural resilience can be enhanced by reducing the uncertainty involved in the process of assessing the condition state of civil infrastructures and carrying out subsequent retrofit work. Over the last decade, deep learning has seen impressive success in traditional pattern recognition applications historically faced with long-time challenges, which motivates the current research in integrating the advancement of deep learning into SHM applications. This dissertation research aims to accomplish the overall goal of establishing a smart SHM technique based on deep learning algorithm, which will achieve automated structural health condition assessment and condition rating prediction for civil infrastructures. A literate review on structural health condition assessment technologies commonly used for civil infrastructures was first conducted to identify the special need of the proposed method. Deep learning algorithms were also reviewed, with a focus on pattern recognition application, especially in the computer vision field in which deep learning algorithms have reported great success in traditionally challenging tasks. Subsequently, a technical procedure is established to incorporate a particular type of deep learning algorithm, termed Convolutional Neural Network which was found behind the many success seen in computer vision applications, into smart SHM technologies. The proposed method was first demonstrated and validated on an SHM application problem that uses image data for structural steel condition assessment. Further study was performed on time series data including vibration data and guided Lamb wave signals for two types of SHM applications - brace damage detection in concentrically braced frame structures or nondestructive evaluation (NDE) of thin plate structures. Additionally, discrete data (neither images nor time series data), such as the bridge condition rating data from National Bridge Inventory (NBI) data repository, was also investigated for the application of bridge condition forecasting. The study results indicated that the proposed method is very promising as a data-driven structural health condition assessment technique for civil infrastructures, based on research findings in the four distinct SHM case studies in this dissertation

    Automatic Skin Cancer Detection in Dermoscopy Images Based on Ensemble Lightweight Deep Learning Network

    Get PDF
    The complex detection background and lesion features make the automatic detection of dermoscopy image lesions face many challenges. The previous solutions mainly focus on using larger and more complex models to improve the accuracy of detection, there is a lack of research on significant intra-class differences and inter-class similarity of lesion features. At the same time, the larger model size also brings challenges to further algorithm application; In this paper, we proposed a lightweight skin cancer recognition model with feature discrimination based on fine-grained classification principle. The propose model includes two common feature extraction modules of lesion classification network and a feature discrimination network. Firstly, two sets of training samples (positive and negative sample pairs) are input into the feature extraction module (Lightweight CNN) of the recognition model. Then, two sets of feature vectors output from the feature extraction module are used to train the two classification networks and feature discrimination networks of the recognition model at the same time, and the model fusion strategy is applied to further improve the performance of the model, the proposed recognition method can extract more discriminative lesion features and improve the recognition performance of the model in a small amount of model parameters; In addition, based on the feature extraction module of the proposed recognition model, U-Net architecture, and migration training strategy, we build a lightweight semantic segmentation model of lesion area of dermoscopy image, which can achieve high precision lesion area segmentation end-to-end without complicated image preprocessing operation; The performance of our approach was appraised through widespread experiments comparative and feature visualization analysis, the outcome indicates that the proposed method has better performance than the start-of-the-art deep learning-based approach on the ISBI 2016 skin lesion analysis towards melanoma detection challenge dataset

    Multimodal Approaches to Computer Vision Problems

    Get PDF
    The goal of computer vision research is to automatically extract high-level information from images and videos. The vast majority of this research focuses specifically on visible light imagery. In this dissertation, we present approaches to computer vision problems that incorporate data obtained from alternative modalities including thermal infrared imagery, near-infrared imagery, and text. We consider approaches where other modalities are used in place of visible imagery as well as approaches that use other modalities to improve the performance of traditional computer vision algorithms. The bulk of this dissertation focuses on Heterogeneous Face Recognition (HFR). HFR is a variant of face recognition where the probe and gallery face images are obtained with different sensing modalities. We also present a method to incorporate text information into human activity recognition algorithms. We first present a kernel task-driven coupled dictionary model to represent the data across multiple domains for thermal infrared HFR. We extend a linear coupled dictionary model to use the kernel method to process the signals in a high dimensional space; this effectively enables the dictionaries to represent the data non-linearly in the original feature space. We further improve the model by making the dictionaries task-driven. This allows us to tune the dictionaries to perform well on the classification task at hand rather than the standard reconstruction task. We show that our algorithms outperform algorithms based on standard coupled dictionaries on three datasets for thermal infrared to visible face recognition. Next, we present a deep learning-based approach to near-infrared (NIR) HFR. Most approaches to HFR involve modeling the relationship between corresponding images from the visible and sensing domains. Due to data constraints, this is typically done at the patch level and/or with shallow models to prevent overfitting. In this approach, rather than modeling local patches or using a simple model, we use a complex, deep model to learn the relationship between the entirety of cross-modal face images. We describe a deep convolutional neural network-based method that leverages a large visible image face dataset to prevent overfitting. We present experimental results on two benchmark data sets showing its effectiveness. Third, we present a model order selection algorithm for deep neural networks. In recent years, deep learning has emerged as a dominant methodology in machine learning. While it has been shown to produce state-of-the-art results for a variety of applications, one aspect of deep networks that has not been extensively researched is how to determine the optimal network structure. This problem is generally solved by ad hoc methods. In this work we address a sub-problem of this task: determining the breadth (number of nodes) of each layer. We show how to use group-sparsity-inducing regularization to automatically select these hyper-parameters. We demonstrate the proposed method by using it to reduce the size of networks while maintaining performance for our NIR HFR deep-learning algorithm. Additionally, we demonstrate the generality of our algorithm by applying it to image classification tasks. Finally, we present a method to improve activity recognition algorithms through the use of multitask learning and information extracted from a large text corpora. Current state-of-the-art deep learning approaches are limited by the size and scope of the data set they use to train the networks. We present a multitask learning approach to expand the training data set. Specifically, we train the neural networks to recognize objects in addition to activities. This allows us to expand our training set with large, publicly available object recognition data sets and thus use deeper, state-of-the-art network architectures. Additionally, when learning about the target activities, the algorithms are limited to the information contained in the training set. It is virtually impossible to capture all variations of the target activities in a training set. In this work, we extract information about the target activities from a large text corpora. We incorporate this information into the training algorithm by using it to select relevant object recognition classes for the multitask learning approach. We present experimental results on a benchmark activity recognition data set showing the effectiveness of our approach

    End-to-End Deep Learning Systems for Scene Understanding, Path Planning and Navigation in Fire Fighter Teams

    Get PDF
    Firefighting is a dynamic activity with many operations occurring simultaneously. Maintaining situational awareness, defined as knowledge of current conditions and activities at the scene, are critical to accurate decision making. Firefighters often carry various sensors in their personal equipment, namely thermal cameras, gas sensors, and microphones. Improved data processing techniques can mine this data more effectively and be used to improve situational awareness at all times thereby improving real-time decision making and minimizing errors in judgment induced by environmental conditions and anxiety levels. This objective of this research employs state of the art Machine Learning (ML) techniques to create an automated system that is capable of real-time object detection and recognition utilizing currently gathered data to achieve improved situational awareness of firefighters on the scene. The algorithms authored effectively exploit the information gathered from the infrared camera by using a trained deep Convolutional Neural Network (CNN) system to identify, classify and track objects of interest. Crucial information is identified and relayed back to firefighters to assist their decision making processes and aid in safely navigating the environment. The ANN-based algorithm we are authoring is sufficient to infer human recognition and posture detection to deduce a victim’s health level to assist in prioritizing victims by need and guide firefighters accordingly. We also employ deep-learning based systems path planning and navigation, path reconstruction, scene segmentation, estimation of firefighter condition, Natural Language Processing for informing firefighter about the scene. We will integrate our search and rescue system with the image recognition system to produce a new search and rescue method that adapts to the changing environment by using Deep Q-learning
    • …
    corecore