348 research outputs found

    Fade Depth Prediction Using Human Presence for Real Life WSN Deployment

    Get PDF
    Current problem in real life WSN deployment is determining fade depth in indoor propagation scenario for link power budget analysis using (fade margin parameter). Due to the fact that human presence impacts the performance of wireless networks, this paper proposes a statistical approach for shadow fading prediction using various real life parameters. Considered parameters within this paper include statistically mapped human presence and the number of people through time compared to the received signal strength. This paper proposes an empirical model fade depth prediction model derived from a comprehensive set of measured data in indoor propagation scenario. It is shown that the measured fade depth has high correlations with the number of people in non-line-of-sight condition, giving a solid foundation for the fade depth prediction model. In line-of-sight conditions this correlations is significantly lower. By using the proposed model in real life deployment scenarios of WSNs, the data loss and power consumption can be reduced by the means of intelligently planning and designing Wireless Sensor Network

    RECEIVE SIGNAL STRENGTH PREDICTION IN THE GSM BAND USING WAVELET DECOMPOSITION

    Get PDF
    In this work, GSM receive signal strength was monitored in an indoor environment. Samples of  GSM receive signal strength was measured on a Mobile Equipment (ME). One-dimensional multilevel wavelet decomposition technique was used to predict the fading phenomenon of the GSM receive signal strength measured. Wavelet prediction revealed that the GSM receive signal strength is attenuated due to the slow fading phenomenon, which fades about 3 times faster than the radio wavelength. The prediction is further validated using probability density functions in terms of Gaussian and Rayleigh distributions. It is observed that, significant part of the signal strength measured is dominated by good signal (- 101 dBm to – 74 dBm) with an average of – 88.8842 dBm and the signal strength followed more of Gaussian than Rayleigh distribution. This confirmed the wavelet prediction.  http://dx.doi.org/10.4314/njt.v36i1.2

    Wavelet Packet Division Multiplexing (WPDM)-Aided Industrial WSNs

    Full text link
    Industrial Internet-of-Things (IIoT) involve multiple groups of sensors, each group sending its observations on a particular phenomenon to a central computing platform over a multiple access channel (MAC). The central platform incorporates a decision fusion center (DFC) that arrives at global decisions regarding each set of phenomena by combining the received local sensor decisions. Owing to the diverse nature of the sensors and heterogeneous nature of the information they report, it becomes extremely challenging for the DFC to denoise the signals and arrive at multiple reliable global decisions regarding multiple phenomena. The industrial environment represents a specific indoor scenario devoid of windows and filled with different noisy electrical and measuring units. In that case, the MAC is modelled as a large-scale shadowed and slowly-faded channel corrupted with a combination of Gaussian and impulsive noise. The primary contribution of this paper is to propose a flexible, robust and highly noise-resilient multi-signal transmission framework based on Wavelet packet division multiplexing (WPDM). The local sensor observations from each group of sensors are waveform coded onto wavelet packet basis functions before reporting them over the MAC. We assume a multi-antenna DFC where the waveform-coded sensor observations can be separated by a bank of linear filters or a correlator receiver, owing to the orthogonality of the received waveforms. At the DFC we formulate and compare fusion rules for fusing received multiple sensor decisions, to arrive at reliable conclusions regarding multiple phenomena. Simulation results show that WPDM-aided wireless sensor network (WSN) for IIoT environments offer higher immunity to noise by more than 10 times over performance without WPDM in terms of probability of false detection

    Twin Delayed DDPG based Dynamic Power Allocation for Mobility in IoRT

    Get PDF
    The internet of robotic things (IoRT) is a modern as well as fast-evolving technology employed in abundant socio-economical aspects which connect user equipment (UE) for communication and data transfer among each other. For ensuring the quality of service (QoS) in IoRT applications, radio resources, for example, transmitting power allocation (PA), interference management, throughput maximization etc., should be efficiently employed and allocated among UE. Traditionally, resource allocation has been formulated using optimization problems, which are then solved using mathematical computer techniques. However, those optimization problems are generally nonconvex as well as nondeterministic polynomial-time hardness (NP-hard). In this paper, one of the most crucial challenges in radio resource management is the emitting power of an antenna called PA, considering that the interfering multiple access channel (IMAC) has been considered. In addition, UE has a natural movement behavior that directly impacts the channel condition between remote radio head (RRH) and UE. Additionally, we have considered two well-known UE mobility models i) random walk and ii) modified Gauss-Markov (GM). As a result, the simulation environment is more realistic and complex. A data-driven as well as model-free continuous action based deep reinforcement learning algorithm called twin delayed deep deterministic policy gradient (TD3) has been proposed that is the combination of policy gradient, actor-critics, as well as double deep Q-learning (DDQL). It optimizes the PA for i) stationary UE, ii) the UE movements according to random walk model, and ii) the UE movement based on the modified GM model. Simulation results show that the proposed TD3 method outperforms model-based techniques like weighted MMSE (WMMSE) and fractional programming (FP) as well as model-free algorithms, for example, deep Q network (DQN) and DDPG in terms of average sum-rate performance

    Research on Cognitive Radio within the Freeband-AAF project

    Get PDF

    Machine learning algorithms for cognitive radio wireless networks

    Get PDF
    In this thesis new methods are presented for achieving spectrum sensing in cognitive radio wireless networks. In particular, supervised, semi-supervised and unsupervised machine learning based spectrum sensing algorithms are developed and various techniques to improve their performance are described. Spectrum sensing problem in multi-antenna cognitive radio networks is considered and a novel eigenvalue based feature is proposed which has the capability to enhance the performance of support vector machines algorithms for signal classification. Furthermore, spectrum sensing under multiple primary users condition is studied and a new re-formulation of the sensing task as a multiple class signal detection problem where each class embeds one or more states is presented. Moreover, the error correcting output codes based multi-class support vector machines algorithms is proposed and investigated for solving the multiple class signal detection problem using two different coding strategies. In addition, the performance of parametric classifiers for spectrum sensing under slow fading channel is studied. To address the attendant performance degradation problem, a Kalman filter based channel estimation technique is proposed for tracking the temporally correlated slow fading channel and updating the decision boundary of the classifiers in real time. Simulation studies are included to assess the performance of the proposed schemes. Finally, techniques for improving the quality of the learning features and improving the detection accuracy of sensing algorithms are studied and a novel beamforming based pre-processing technique is presented for feature realization in multi-antenna cognitive radio systems. Furthermore, using the beamformer derived features, new algorithms are developed for multiple hypothesis testing facilitating joint spatio-temporal spectrum sensing. The key performance metrics of the classifiers are evaluated to demonstrate the superiority of the proposed methods in comparison with previously proposed alternatives
    corecore