40 research outputs found
Prediction-Based Channel Selection Prediction in Mobile Cognitive Radio Network
The emerging 5G wireless communications enabled diverse multimedia applications and smart devices in the network. It promises very high mobile traffic data rates, quality of service as in very low latency and improvement in user’s perceived quality of experience compared to current 4G wireless network. This encourages the increasing demand of significant bandwidth which results a significant urge of efficient spectrum utilization. In this paper, modelling, performance analysis and optimization of future channel selection for cognitive radio network by jointly exploiting both CR mobility and primary user activity to provide efficient spectrum access is studied. The modelling and prediction method is implemented by using Hidden Markov Model algorithm. The movement of CR in wireless network yields location-varying spectrum opportunities. The current approaches in most literatures which only depend on reactive selection spectrum opportunities result of inefficient channel usages. Moreover, conventional random selection method tends to observe a higher handoff and operation delays in network performance. This inefficiency can cause continuous transmission interruptions leading to the degradation of advance wireless services. This work goal is to improve the performance of CR in terms number of handoffs and operation delays. We perform simulation on our prediction strategy with a commonly used random sensing method with and without location. Through simulations, it is shown that the proposed prediction and learning strategy can obtain significant improvements in number of handoffs and operation delays performance parameters. It is also shown that future CR location is beneficial in increasing mobile CR performance. This study also shows that the number of primary user in the network and the PU protection range affect the performance of mobile CR channel selection for all methods
Breast Cancer Classification: Features Investigation using Machine Learning Approaches
Breast cancer is the second most common cancer after lung cancer and one of the main causes of death worldwide. Women have a higher risk of breast cancer as compared to men. Thus, one of the early diagnosis with an accurate and reliable system is critical in breast cancer treatment. Machine learning techniques are well known and popular among researchers, especially for classification and prediction. An investigation was conducted to evaluate the performance of breast cancer classification for malignant tumors and benign tumors using various machine learning techniques, namely k-Nearest Neighbors (k-NN), Random Forest, and Support Vector Machine (SVM) and ensemble techniques to compute the prediction of the breast cancer survival by implementing 10-fold cross validation. This study used a dataset obtained from Wisconsin Diagnostic Breast Cancer (WDBC) with 23 selected features measured from 569 patients, from which 212 patients have malignant tumors and 357 patients have benign tumors. The analysis was performed to investigate the feature of the tumors based on its mean, standard error, and worst. Each feature has ten properties which are radius, texture, perimeter, area, smoothness, compactness, concavity, concave, symmetry and fractal dimensions. The selection of features was considered a significant influence to the breast cancer. The analysis is compared and evaluated with thirty features to determine the features used for breast cancer classification. The result shown AdaBoost has obtained the highest accuracy for thirty features at 98.95%, ten features of mean at 98.07%, and ten features of worst at 98.77% with a lowest error rate. Additionally, the proposed methods are classified using 2-fold, 3-fold, and 5-fold cross validation to meet the best accuracy rate. Comparison results between all methods show that AdaBoost ensemble methods gave the highest accuracy at 98.77% for 10-fold cross validation, while 2-fold and 3-fold cross validation at 98.41% and 98.24%, respectively. Nevertheless, the result with 5-fold cross validation shows SVM produced the best accuracy rate at 98.60% with the lowest error rate
Joint Source Channel Decoding Exploiting 2D Source Correlation with Parameter Estimation for Image Transmission over Rayleigh Fading Channels
This paper investigates the performance of a 2- Dimensional (2D) Joint Source Channel Coding (JSCC) system assisted with parameter estimation for 2D image transmission over an Additive White Gaussian Noise (AWGN) channel and a Rayleigh fading channel. Baum-Welsh Algorithm (BWA)  is employed in the proposed 2D JSCC system to estimate the source correlation statistics during channel decoding. The source correlation is then exploited during channel decoding using a Modified Bahl-Cocke-Jelinek-Raviv (BCJR) algorithm. The performance of the 2D JSCC system with the BWA-based parameter estimation technique (2D-JSCC-PET1) is evaluated via image transmission simulations. Two images, each exhibits strong and weak source correlation are considered in the evaluation by measuring the Peak Signal Noise Ratio of the decoded images at the receiver. The proposed 2D-JSCC-PET1 system is compared with various benchmark systems. Simulation results reveal that the 2D-JSCC-PET1 system outperforms the other benchmark systems (performance gain of 4.23 dB over the 2D-JSCC-PET2 system and 6.10 dB over the 2D JSCC system). The proposed system also can perform very close to the ideal 2D JSCC system relying on the assumption of perfect source correlation knowledge at the receiver that shown only 0.88 dB difference in performance gain
PAPR reduction techniques in generalized inverse discrete fourier transform non-orthogonal frequency division multiplexing system
A promising system of Generalized Inverse Discrete Fourier Transform Non-Orthogonal Frequency Division Multiplexing (GIDFT n-OFDM) system can fulfil the requirement of supporting higher data rate in Fifth Generation (5G) technology. However, this system experience High Peak to Average Power Ratio (PAPR) due to massive number of subcarriers signal is transmitted. In this paper, three types of usual PAPR reduction techniques were applied in GIDFT n-OFDM system which are Clipping, Partial transmit Transform (PTS) and Selective Mapping (SLM). The system performance is compared and evaluated using Complementary Cumulative Distribution Function (CCDF) plot. Simulation results show that SLM technique give significant reduction of PAPR 9 dB of the original performance
COVID-19: Symptoms Clustering and Severity Classification Using Machine Learning Approach
COVID-19 is an extremely contagious illness that causes illnesses varying from either the common cold to more chronic illnesses or even death. The constant mutation of a new variant of COVID-19 makes it important to identify the symptom of COVID-19 in order to contain the infection. The use of clustering and classification in machine learning is in mainstream use in different aspects of research, especially in recent years to generate useful knowledge on COVID-19 outbreak. Many researchers have shared their COVID-19 data on public database and a lot of studies have been carried out. However, the merit of the dataset is unknown and analysis need to be carried by the researchers to check on its reliability. The dataset that is used in this work was sourced from the Kaggle website. The data was obtained through a survey collected from participants of various gender and age who had been to at least ten countries. There are four levels of severity based on the COVID-19 symptom, which was developed in accordance to World Health Organization (WHO) and the Indian Ministry of Health and Family Welfare recommendations. Â This paper presented an inquiry on the dataset utilising supervised and unsupervised machine learning approaches in order to better comprehend the dataset. In this study, the analysis of the severity group based on the COVID-19 symptoms using supervised learning techniques employed a total of seven classifiers, namely the K-NN, Linear SVM, Naive Bayes, Decision Tree (J48), Ada Boost, Bagging, and Stacking. For the unsupervised learning techniques, the clustering algorithm utilized in this work are Simple K-Means and Expectation-Maximization. From the result obtained from both supervised and unsupervised learning techniques, we observed that the result analysis yielded relatively poor classification and clustering results. The findings for the dataset analysed in this study do not appear to be providing the correct result for the symptoms categorized against the severity level which raises concerns about the validity and reliability of the dataset
Exploiting 2-Dimensional Source Correlation in Channel Decoding with Parameter Estimation
Traditionally, it is assumed that source coding is perfect and therefore, the redundancy of the source encoded bit-stream is zero. However, in reality, this is not the case as the existing source encoders are imperfect and yield residual redundancy at the output. The residual redundancy can be exploited by using Joint Source Channel Coding (JSCC) with Markov chain as the source. In several studies, the statistical knowledge of the sources has been assumed to be perfectly available at the receiver. Although the result was better in terms of the BER performance, practically, the source correlation knowledge were not always available at the receiver and thus, this could affect the reliability of the outcome. The source correlation on all rows and columns of the 2D sources were well exploited by using a modified Bahl-Cocke-Jelinek-Raviv (BCJR) algorithm in the decoder. A parameter estimation technique was used jointly with the decoder to estimate the source correlation knowledge. Hence, this research aims to investigate the parameter estimation for 2D JSCC system which reflects a practical scenario where the source correlation knowledge are not always available. We compare the performance of the proposed joint decoding and estimation technique with the ideal 2D JSCC system with perfect knowledge of the source correlation knowledge. Simulation results reveal that our proposed coding scheme performs very close to the ideal 2D JSCC system
COVID-19: Symptoms Clustering and Severity Classification Using Machine Learning Approach
COVID-19 is an extremely contagious illness that causes illnesses varying from either the common cold to more chronic illnesses or even death. The constant mutation of a new variant of COVID-19 makes it important to identify the symptom of COVID-19 in order to contain the infection. The use of clustering and classification in machine learning is in mainstream use in different aspects of research, especially in recent years to generate useful knowledge on COVID-19 outbreak. Many researchers have shared their COVID-19 data on public database and a lot of studies have been carried out. However, the merit of the dataset is unknown and analysis need to be carried by the researchers to check on its reliability. The dataset that is used in this work was sourced from the Kaggle website. The data was obtained through a survey collected from participants of various gender and age who had been to at least ten countries. There are four levels of severity based on the COVID-19 symptom, which was developed in accordance to World Health Organization (WHO) and the Indian Ministry of Health and Family Welfare recommendations. Â This paper presented an inquiry on the dataset utilising supervised and unsupervised machine learning approaches in order to better comprehend the dataset. In this study, the analysis of the severity group based on the COVID-19 symptoms using supervised learning techniques employed a total of seven classifiers, namely the K-NN, Linear SVM, Naive Bayes, Decision Tree (J48), Ada Boost, Bagging, and Stacking. For the unsupervised learning techniques, the clustering algorithm utilized in this work are Simple K-Means and Expectation-Maximization. From the result obtained from both supervised and unsupervised learning techniques, we observed that the result analysis yielded relatively poor classification and clustering results. The findings for the dataset analysed in this study do not appear to be providing the correct result for the symptoms categorized against the severity level which raises concerns about the validity and reliability of the dataset
Internet of Things: A Monitoring and Control System for Rockmelon Farming
This paper describes the internet of things application in agriculture production especially for rock melon farming. In order to boost agricultural production on a commercial basis, a more systematic approach should be developed and organized to be adopted by operators to increase production and income. This work combines fertilization and irrigation in one system under protective structures to ensure that high-quality plant production and an alternative to conventional cropping systems. In addition, the use of technology for online monitoring and control is an improvement to the system in parallel with the rapid development of information technology today. The proposed system is focusing on automation, wireless, data analytics and simplified design with minimal and scalable skid size for rock melon in the Klang valley. Other than that, a monitoring and control system were developed besides, applied Internet of Thing (IoT) platform with additional user-friendly programmable farming routine Human-machine interfacing (HMI). HMI is a software interface that is capable to conduct the systems directly via autonomous cloud control. It provides time management and contributes to a more efficient workforce. The aim to develop a monitoring and controlling system using IoT for rock melon in this study was achieved. The rock melon harvesting succeeds according to the scheduled routine and the quality of yields also satisfied
Mobile communication (2G, 3G and 4G) and future interest of 5G in Pakistan: A review
The use of mobile communication is growing radically with every passing year. The new reality is the fifth generation (5G) of mobile communication technology. 5G requires expensive infrastructural adjustment and upgradation. Currently, Pakistan has one of the most significant numbers of biometrically verified mobile users. However, at the same time, the country lags incredibly in the field of mobile internet adoption, with just half of the mobile device owners avail broadband subscription. It is a viable market with a large segment yet to be tapped. With the advancing progression in Pakistan towards the internet of things (IoT) connectivity, i.e., solar-powered home solutions, smart city projects, and on-board diagnostics (OBD), the urgency for speed, bandwidth and reliability are on the rise. In this paper, Pakistan's prevalent mobile communication networks, i.e., second, third and fourth generation (2G, 3G and 4G), were analyzed and examined in light of the country's demographics and challenges. The future of 5G in Pakistan was also discussed. The study revealed that non-infrastructural barriers influence the low adoption rate, which is the main reason behind the spectrum utilization gap, i.e., the use of 3G, and the 4G spectrum is minimal