2,694 research outputs found
Markov Decision Processes with Applications in Wireless Sensor Networks: A Survey
Wireless sensor networks (WSNs) consist of autonomous and resource-limited
devices. The devices cooperate to monitor one or more physical phenomena within
an area of interest. WSNs operate as stochastic systems because of randomness
in the monitored environments. For long service time and low maintenance cost,
WSNs require adaptive and robust methods to address data exchange, topology
formulation, resource and power optimization, sensing coverage and object
detection, and security challenges. In these problems, sensor nodes are to make
optimized decisions from a set of accessible strategies to achieve design
goals. This survey reviews numerous applications of the Markov decision process
(MDP) framework, a powerful decision-making tool to develop adaptive algorithms
and protocols for WSNs. Furthermore, various solution methods are discussed and
compared to serve as a guide for using MDPs in WSNs
Resource Allocation in Wireless Networks with RF Energy Harvesting and Transfer
Radio frequency (RF) energy harvesting and transfer techniques have recently
become alternative methods to power the next generation of wireless networks.
As this emerging technology enables proactive replenishment of wireless
devices, it is advantageous in supporting applications with quality-of-service
(QoS) requirement. This article focuses on the resource allocation issues in
wireless networks with RF energy harvesting capability, referred to as RF
energy harvesting networks (RF-EHNs). First, we present an overview of the
RF-EHNs, followed by a review of a variety of issues regarding resource
allocation. Then, we present a case study of designing in the receiver
operation policy, which is of paramount importance in the RF-EHNs. We focus on
QoS support and service differentiation, which have not been addressed by
previous literatures. Furthermore, we outline some open research directions.Comment: To appear in IEEE Networ
- …