3,323 research outputs found

    A Channel Ranking And Selection Scheme Based On Channel Occupancy And SNR For Cognitive Radio Systems

    Get PDF
    Wireless networks and information traffic have grown exponentially over the last decade. Consequently, an increase in demand for radio spectrum frequency bandwidth has resulted. Recent studies have shown that with the current fixed spectrum allocation (FSA), radio frequency band utilization ranges from 15% to 85%. Therefore, there are spectrum holes that are not utilized all the time by the licensed users, and, thus the radio spectrum is inefficiently exploited. To solve the problem of scarcity and inefficient utilization of the spectrum resources, dynamic spectrum access has been proposed as a solution to enable sharing and using available frequency channels. With dynamic spectrum allocation (DSA), unlicensed users can access and use licensed, available channels when primary users are not transmitting. Cognitive Radio technology is one of the next generation technologies that will allow efficient utilization of spectrum resources by enabling DSA. However, dynamic spectrum allocation by a cognitive radio system comes with the challenges of accurately detecting and selecting the best channel based on the channelâs availability and quality of service. Therefore, the spectrum sensing and analysis processes of a cognitive radio system are essential to make accurate decisions. Different spectrum sensing techniques and channel selection schemes have been proposed. However, these techniques only consider the spectrum occupancy rate for selecting the best channel, which can lead to erroneous decisions. Other communication parameters, such as the Signal-to-Noise Ratio (SNR) should also be taken into account. Therefore, the spectrum decision-making process of a cognitive radio system must use techniques that consider spectrum occupancy and channel quality metrics to rank channels and select the best option. This thesis aims to develop a utility function based on spectrum occupancy and SNR measurements to model and rank the sensed channels. An evolutionary algorithm-based SNR estimation technique was developed, which enables adaptively varying key parameters of the existing Eigenvalue-based blind SNR estimation technique. The performance of the improved technique is compared to the existing technique. Results show the evolutionary algorithm-based estimation performing better than the existing technique. The utility-based channel ranking technique was developed by first defining channel utility function that takes into account SNR and spectrum occupancy. Different mathematical functions were investigated to appropriately model the utility of SNR and spectrum occupancy rate. A ranking table is provided with the utility values of the sensed channels and compared with the usual occupancy rate based channel ranking. According to the results, utility-based channel ranking provides a better scope of making an informed decision by considering both channel occupancy rate and SNR. In addition, the efficiency of several noise cancellation techniques was investigated. These techniques can be employed to get rid of the impact of noise on the received or sensed signals during spectrum sensing process of a cognitive radio system. Performance evaluation of these techniques was done using simulations and the results show that the evolutionary algorithm-based noise cancellation techniques, particle swarm optimization and genetic algorithm perform better than the regular gradient descent based technique, which is the least-mean-square algorithm

    READUP BUILDUP. Thync - instant α-readings

    Get PDF

    Benchmarking Particle Filter Algorithms for Efficient Velodyne-Based Vehicle Localization

    Get PDF
    Keeping a vehicle well-localized within a prebuilt-map is at the core of any autonomous vehicle navigation system. In this work, we show that both standard SIR sampling and rejection-based optimal sampling are suitable for efficient (10 to 20 ms) real-time pose tracking without feature detection that is using raw point clouds from a 3D LiDAR. Motivated by the large amount of information captured by these sensors, we perform a systematic statistical analysis of how many points are actually required to reach an optimal ratio between efficiency and positioning accuracy. Furthermore, initialization from adverse conditions, e.g., poor GPS signal in urban canyons, we also identify the optimal particle filter settings required to ensure convergence. Our findings include that a decimation factor between 100 and 200 on incoming point clouds provides a large savings in computational cost with a negligible loss in localization accuracy for a VLP-16 scanner. Furthermore, an initial density of ∼2 particles/m 2 is required to achieve 100% convergence success for large-scale (∼100,000 m 2 ), outdoor global localization without any additional hint from GPS or magnetic field sensors. All implementations have been released as open-source software

    A Bayesian fusion model for space-time reconstruction of finely resolved velocities in turbulent flows from low resolution measurements

    Full text link
    The study of turbulent flows calls for measurements with high resolution both in space and in time. We propose a new approach to reconstruct High-Temporal-High-Spatial resolution velocity fields by combining two sources of information that are well-resolved either in space or in time, the Low-Temporal-High-Spatial (LTHS) and the High-Temporal-Low-Spatial (HTLS) resolution measurements. In the framework of co-conception between sensing and data post-processing, this work extensively investigates a Bayesian reconstruction approach using a simulated database. A Bayesian fusion model is developed to solve the inverse problem of data reconstruction. The model uses a Maximum A Posteriori estimate, which yields the most probable field knowing the measurements. The DNS of a wall-bounded turbulent flow at moderate Reynolds number is used to validate and assess the performances of the present approach. Low resolution measurements are subsampled in time and space from the fully resolved data. Reconstructed velocities are compared to the reference DNS to estimate the reconstruction errors. The model is compared to other conventional methods such as Linear Stochastic Estimation and cubic spline interpolation. Results show the superior accuracy of the proposed method in all configurations. Further investigations of model performances on various range of scales demonstrate its robustness. Numerical experiments also permit to estimate the expected maximum information level corresponding to limitations of experimental instruments.Comment: 15 pages, 6 figure
    • …
    corecore