273 research outputs found
Power and Time Slot Allocation in Cognitive Relay Networks Using Particle Swarm Optimization
The two main problems in cognitive radio networks are power and time slot allocation problems which require a precise analysis and guarantee the quality of service in both the primary and secondary users. In this paper, these two problems are considered and a method is proposed to solve the resulting optimization problem. Our proposed method provides an improved performance in solving the constrained nonlinear multiobject optimization for the power control and beamforming in order to reach the maximum capacity and proper adaption of time slots, and as a result a new scheme for joint power and time slot allocation in cognitive relay networks is proposed. We adopt space diversity access as the secondary users access scheme and divide the time between multiple secondary users according to their contribution to primary user's transmission. Helping primary users provides more opportunities for secondary users to access the channel since the primary users can release the channel sooner. In contrast, primary network leases portion of channel access time to the secondary users for their transmission using particle swarm optimization (PSO). Numerical studies show good performance of the proposed scheme with a dynamic cost function in a nonstationary environment
Recommended from our members
Capacity Enhancement Approaches for Long Term Evolution networks: Capacity Enhancement-Inspired Self-Organized Networking to Enhance Capacity and Fairness of Traffic in Long Term Evolution Networks by Utilising Dynamic Mobile Base-Stations
The long-term evolution (LTE) network has been proposed to provide better network capacity than the earlier 3G network. Driven by the market, the conventional LTE (3G) network standard could not achieve the expectations of the international mobile telecommunications advanced (IMT-Advanced) standard. To satisfy this gap, the LTE-Advanced was introduced with additional network functionalities to meet up with the IMT-Advanced Standard. In addition, due to the need to minimize operational expenditure (OPEX) and reduce human interventions, the wireless cellular networks are required to be self-aware, self-reconfigurable, self-adaptive and smart. An example of such network involves transceiver base stations (BTSs) within a self-organizing network (SON).
Besides these great breakthroughs, the conventional LTE and LTE-Advanced networks have not been designed with the intelligence of scalable capacity output especially in sudden demographic changes, namely during events of football, malls, worship centres or during religious and cultural festivals. Since most of these events cannot be predicted, modern cellular networks must be scalable in terms of capacity and coverage in such unpredictable demographic surge. Thus, the use of dynamic BTSs is proposed to be used in modern and future cellular networks for crowd and demographic change managements.
Dynamic BTSs are complements of the capability of SONs to search, determine and deploy less crowded/idle BTSs to densely crowded cells for scalable capacity management. The mobile BTSs will discover areas of dark coverages and fill-up the gap in terms of providing cellular services. The proposed network relieves the LTE network from overloading thus reducing packet loss, delay and improves fair load sharing.
In order to trail the best (least) path, a bio-inspired optimization algorithm based on swarm-particle optimization is proposed over the dynamic BTS network. It uses the ant-colony optimization algorithm (ACOA) to find the least path. A comparison between an optimized path and the un-optimized path showed huge gain in terms of delay, fair load sharing and the percentage of packet loss
Machine Learning Meets Communication Networks: Current Trends and Future Challenges
The growing network density and unprecedented increase in network traffic, caused by the massively expanding number of connected devices and online services, require intelligent network operations. Machine Learning (ML) has been applied in this regard in different types of networks and networking technologies to meet the requirements of future communicating devices and services. In this article, we provide a detailed account of current research on the application of ML in communication networks and shed light on future research challenges. Research on the application of ML in communication networks is described in: i) the three layers, i.e., physical, access, and network layers; and ii) novel computing and networking concepts such as Multi-access Edge Computing (MEC), Software Defined Networking (SDN), Network Functions Virtualization (NFV), and a brief overview of ML-based network security. Important future research challenges are identified and presented to help stir further research in key areas in this direction
Contextual Beamforming: Exploiting Location and AI for Enhanced Wireless Telecommunication Performance
The pervasive nature of wireless telecommunication has made it the foundation
for mainstream technologies like automation, smart vehicles, virtual reality,
and unmanned aerial vehicles. As these technologies experience widespread
adoption in our daily lives, ensuring the reliable performance of cellular
networks in mobile scenarios has become a paramount challenge. Beamforming, an
integral component of modern mobile networks, enables spatial selectivity and
improves network quality. However, many beamforming techniques are iterative,
introducing unwanted latency to the system. In recent times, there has been a
growing interest in leveraging mobile users' location information to expedite
beamforming processes. This paper explores the concept of contextual
beamforming, discussing its advantages, disadvantages and implications.
Notably, the study presents an impressive 53% improvement in signal-to-noise
ratio (SNR) by implementing the adaptive beamforming (MRT) algorithm compared
to scenarios without beamforming. It further elucidates how MRT contributes to
contextual beamforming. The importance of localization in implementing
contextual beamforming is also examined. Additionally, the paper delves into
the use of artificial intelligence schemes, including machine learning and deep
learning, in implementing contextual beamforming techniques that leverage user
location information. Based on the comprehensive review, the results suggest
that the combination of MRT and Zero forcing (ZF) techniques, alongside deep
neural networks (DNN) employing Bayesian Optimization (BO), represents the most
promising approach for contextual beamforming. Furthermore, the study discusses
the future potential of programmable switches, such as Tofino, in enabling
location-aware beamforming
- …