19 research outputs found

    Model Based Ceramic tile inspection using Discrete Wavelet Transform and Euclidean Distance

    Full text link
    Visual inspection of industrial products is used to determine the control quality for these products. This paper deals with the problem of visual inspection of ceramic tiles industry using Wavelet Transform. The third level the coefficients of two dimensions Haar Discrete Wavelet Transform (HDWT) is used in this paper to process the images and feature extraction. The proposed algorithm consists of two main phases. The first phase is to compute the wavelet transform for an image free of defects which known as reference image, and the image to be inspected which known as test image. The second phase is used to decide whether the tested image is defected or not using the Euclidean distance similarity measure. The experimentation results of the proposed algorithm give 97% for correct detection of ceramic defects.Comment: Pages IEEE format, International Journal of Computer Science and Information Security, IJCSIS, Vol. 7 No. 2, February 2010, USA. ISSN 1947 5500, http://sites.google.com/site/ijcsis

    m-ary Balanced Codes With Parallel Decoding

    Get PDF

    The Role of 3D CT Imaging in the Accurate Diagnosis of Lung Function in Coronavirus Patients

    Get PDF
    Early grading of coronavirus disease 2019 (COVID-19), as well as ventilator support machines, are prime ways to help the world fight this virus and reduce the mortality rate. To reduce the burden on physicians, we developed an automatic Computer-Aided Diagnostic (CAD) system to grade COVID-19 from Computed Tomography (CT) images. This system segments the lung region from chest CT scans using an unsupervised approach based on an appearance model, followed by 3D rotation invariant Markov–Gibbs Random Field (MGRF)-based morphological constraints. This system analyzes the segmented lung and generates precise, analytical imaging markers by estimating the MGRF-based analytical potentials. Three Gibbs energy markers were extracted from each CT scan by tuning the MGRF parameters on each lesion separately. The latter were healthy/mild, moderate, and severe lesions. To represent these markers more reliably, a Cumulative Distribution Function (CDF) was generated, then statistical markers were extracted from it, namely, 10th through 90th CDF percentiles with 10% increments. Subsequently, the three extracted markers were combined together and fed into a backpropagation neural network to make the diagnosis. The developed system was assessed on 76 COVID-19-infected patients using two metrics, namely, accuracy and Kappa. In this paper, the proposed system was trained and tested by three approaches. In the first approach, the MGRF model was trained and tested on the lungs. This approach achieved 95.83% accuracy and 93.39% kappa. In the second approach, we trained the MGRF model on the lesions and tested it on the lungs. This approach achieved 91.67% accuracy and 86.67% kappa. Finally, we trained and tested the MGRF model on lesions. It achieved 100% accuracy and 100% kappa. The results reported in this paper show the ability of the developed system to accurately grade COVID-19 lesions compared to other machine learning classifiers, such as k-Nearest Neighbor (KNN), decision tree, naïve Bayes, and random forest

    An Efficiency Boost for Genetic Algorithms: Initializing the GA with the Iterative Approximate Method for Optimizing the Traveling Salesman Problem—Experimental Insights

    No full text
    The genetic algorithm (GA) is a well-known metaheuristic approach for dealing with complex problems with a wide search space. In genetic algorithms (GAs), the quality of individuals in the initial population is important in determining the final optimal solution. The classic GA using the random population seeding technique is effective and straightforward, but the generated population may contain individuals with low fitness, delaying convergence to the best solution. On the other side, heuristic population seeding strategies provide the advantages of producing individuals with high fitness and encouraging rapid convergence to the optimal solution. Using background information on the problem being solved, researchers have developed several population seeding approaches. In this paper, to enhance the genetic algorithm efficiency, we propose a new method for the initial population seeding based on a greedy approach. The proposed method starts by adding four extreme cities to the route, creating a loop, and then adding each city to the route through a greedy strategy that calculates the cost of adding every city to different locations along the route. This method identifies the best position to place the city as well as the best city to add. Employing local constant permutations improves the resultant route even more. Together with the suggested approach, we examine the GA’s effectiveness while employing conventional population seeding methods like nearest-neighbor, regression-based, and random seeding. Utilizing some of the well-known Traveling Salesman Problem (TSP) examples from the TSPLIB, the standard library for TSPs, tests were conducted. In terms of the error rate, average convergence, and time, the experimental results demonstrate that the GA that employs the suggested population seeding technique performs better than other GAs that use conventional population seeding strategies

    Smart Attacks Learning Machine Advisor System for Protecting Smart Cities from Smart Threats

    No full text
    The extensive use of Internet of Things (IoT) technology has recently enabled the development of smart cities. Smart cities operate in real-time to improve metropolitan areas’ comfort and efficiency. Sensors in these IoT devices are immediately linked to enormous servers, creating smart city traffic flow. This flow is rapidly increasing and is creating new cybersecurity concerns. Malicious attackers increasingly target essential infrastructure such as electricity transmission and other vital infrastructures. Software-Defined Networking (SDN) is a resilient connectivity technology utilized to address security concerns more efficiently. The controller, which oversees the flows of each appropriate forwarding unit in the SDN architecture, is the most critical component. The controller’s flow statistics are thought to provide relevant information for building an Intrusion Detection System (IDS). As a result, we propose a five-level classification approach based on SDN’s flow statistics to develop a Smart Attacks Learning Machine Advisor (SALMA) system for detecting intrusions and for protecting smart cities from smart threats. We use the Extreme Learning Machine (ELM) technique at all levels. The proposed system was implemented on the NSL-KDD and KDDCUP99 benchmark datasets, and achieved 95% and 99.2%, respectively. As a result, our approach provides an effective method for detecting intrusions in SDNs

    Improving the efficiency of RMSProp optimizer by utilizing Nestrove in deep learning

    No full text
    Abstract There are several methods that have been discovered to improve the performance of Deep Learning (DL). Many of these methods reached the best performance of their models by tuning several parameters such as Transfer Learning, Data augmentation, Dropout, and Batch Normalization, while other selects the best optimizer and the best architecture for their model. This paper is mainly concerned with the optimization algorithms in DL. It proposes a modified version of Root Mean Squared Propagation (RMSProp) algorithm, called NRMSProp, to improve the speed of convergence, and to find the minimum of the loss function quicker than the original RMSProp optimizer. Moreover, NRMSProp takes the original algorithm, RMSProp, a step further by using the advantages of Nesterov Accelerated Gradient (NAG). It also takes in consideration the direction of the gradient at the next step, with respect to the history of the previous gradients, and adapts the value of the learning rate. As a result, this modification helps NRMSProp to convergence quicker than the original RMSProp, without any increase in the complexity of the RMSProp. In this work, many experiments had been conducted to evaluate the performance of NRMSProp with performing several tests with deep Convolution Neural Networks (CNNs) using different datasets on RMSProp, Adam, and NRMSProp optimizers. The experimental results showed that NRMSProp has achieved effective performance, and accuracy up to 0.97 in most cases, in comparison to RMSProp and Adam optimizers, without any increase in the complexity of the algorithm and with fine amount of memory and time

    STATEFUL LAYERED CHAIN MODEL TO IMPROVE THE SCALABILITY OF BITCOIN

    No full text
    Bitcoin becomes the focus of scientific research in the modern era. Blockchain is the underlying technology of Bitcoin because of its decentralization, transparency, trust-less, and immutability features. However, blockchain can be considered the cause of Bitcoin scalability issues, especially storage. Nodes in the Bitcoin network need to store the full blockchain to validate transactions. Over time, the blockchain size will be bulky. So, the full nodes will prefer to leave the network. This leads to the blockchain being centralized and trusted, and the security will be adversely affected. This paper proposes a Stateful Layered Chain Model based on storing accounts' balances to reduce the Bitcoin blockchain size. This model changes the structure of the traditional blockchain from blocks to layers. The experimental results demonstrated that the proposed model reduces the blockchain size by about 50.6 %. Implicitly, the transaction throughput can also be doubled. [JJCIT 2023; 9(2.000): 137-153

    Fine tuning deep learning models for breast tumor classification

    No full text
    Abstract This paper proposes an approach to enhance the differentiation task between benign and malignant Breast Tumors (BT) using histopathology images from the BreakHis dataset. The main stages involve preprocessing, which encompasses image resizing, data partitioning (training and testing sets), followed by data augmentation techniques. Both feature extraction and classification tasks are employed by a Custom CNN. The experimental results show that the proposed approach using the Custom CNN model exhibits better performance with an accuracy of 84% than applying the same approach using other pretrained models, including MobileNetV3, EfficientNetB0, Vgg16, and ResNet50V2, that present relatively lower accuracies, ranging from 74 to 82%; these four models are used as both feature extractors and classifiers. To increase the accuracy and other performance metrics, Grey Wolf Optimization (GWO), and Modified Gorilla Troops Optimization (MGTO) metaheuristic optimizers are applied to each model separately for hyperparameter tuning. In this case, the experimental results show that the Custom CNN model, refined with MGTO optimization, reaches an exceptional accuracy of 93.13% in just 10 iterations, outperforming the other state-of-the-art methods, and the other four used pretrained models based on the BreakHis dataset

    Wireless body area sensor networks based human activity recognition using deep learning

    No full text
    Abstract In the healthcare sector, the health status and biological, and physical activity of the patient are monitored among different sensors that collect the required information about these activities using Wireless body area network (WBAN) architecture. Sensor-based human activity recognition (HAR), which offers remarkable qualities of ease and privacy, has drawn increasing attention from researchers with the growth of the Internet of Things (IoT) and wearable technology. Deep learning has the ability to extract high-dimensional information automatically, making end-to-end learning. The most significant obstacles to computer vision, particularly convolutional neural networks (CNNs), are the effect of the environment background, camera shielding, and other variables. This paper aims to propose and develop a new HAR system in WBAN dependence on the Gramian angular field (GAF) and DenseNet. Once the necessary signals are obtained, the input signals undergo pre-processing through artifact removal and median filtering. In the initial stage, the time series data captured by the sensors undergoes a conversion process, transforming it into 2-dimensional images by using the GAF algorithm. Then, DenseNet automatically makes the processes and integrates the data collected from diverse sensors. The experiment results show that the proposed method achieves the best outcomes in which it achieves 97.83% accuracy, 97.83% F-measure, and 97.64 Matthews correlation coefficient (MCC)

    Ceramic Tile Inspection Using a Hybrid Model of Discrete Wavelet Transform and Gray Level Co-Occurrence Matrix and Artificial Neural Networks

    No full text
    Visual inspection of industrial products is used to determine the control quality for different products. This paper deals with the problem of visual inspection of ceramic tiles industry using a hybrid model of 2-D Discrete Wavelet Transform (DWT) with Haar as a mother wavelet, Gray-Level Co-occurrence Matrices (GLCM) and Artificial Neural Networks (ANN). The proposed algorithm consists of two main phases, the first phase is feature extraction and it is carried out using the first level coefficients of two dimensions DWT and GLCM. The second phase is classification which is carried out using two different supervised ANN classifiers; Multilayer Perceptron (MLP) and Radial Basis Function (RBF) networks. The experimental results of the proposed algorithm give good results in defect detection especially with RBF (average 94 %) rather than MLP (average 91 %)
    corecore