1,687 research outputs found
Game theory framework for MAC parameter optimization in energy-delay constrained sensor networks
Optimizing energy consumption and end-to-end (e2e) packet delay in energy-constrained, delay-sensitive wireless sensor networks is a conflicting multiobjective optimization problem. We investigate the problem from a game theory perspective, where the two optimization objectives are considered as game players. The cost model of each player is mapped through a generalized optimization framework onto protocol-specific MAC parameters. From the optimization framework, a game is first defined by the Nash bargaining solution (NBS) to assure energy consumption and e2e delay balancing. Secondy, the Kalai-Smorodinsky bargaining solution (KSBS) is used to find an equal proportion of gain between players. Both methods offer a bargaining solution to the duty-cycle MAC protocol under different axioms. As a result, given the two performance requirements (i.e., the maximum latency tolerated by the application and the initial energy budget of nodes), the proposed framework allows to set tunable system parameters to reach a fair equilibrium point that dually minimizes the system latency and energy consumption. For illustration, this formulation is applied to six state-of-the-art wireless sensor network (WSN) MAC protocols: B-MAC, X-MAC, RI-MAC, SMAC, DMAC, and LMAC. The article shows the effectiveness and scalability of such a framework in optimizing protocol parameters that achieve a fair energy-delay performance trade-off under the application requirements
Recommended from our members
Analysis of a class of distributed queues with application
Recently we have developed a class of media access control algorithms for different types of Local Area Networks. A common feature of these LAN algorithms is that they represent various strategies by which the processors in the LAN can simulate the availability of a centralized packet transport facility, but whose service incorporates a particular type of change over time known as 'moving sever' overhead. First we describe the operation of moving server systems in general, for both First-Come - First-Served and Head-of-the-Line orders of service, together with an approach for their delay analysis in which we transform the moving server queueing system into a conventional queueing system having proportional waiting times. Then we describe how the various LAN algorithms may be obtained from the ideal moving server system, and how a significant component of their performance characteristics is determined by the performance characteristics of that ideal system. Finally, we evaluate the compatibility of such LAN algorithms with separable queueing network models of distributed systems by computing the interdeparture time distribution for M/M/1 in the presence of moving server overhead. Although it is not exponential, except in the limits of low server utilization or low overhead, the interdeparture time distribution is a weighted sum of exponential terms with a coefficient of variation not much smaller than unity. Thus, we conjecture that a service centre with moving server overhead could be used to represent one of these LAN algorithms in a product form queueing network model of a distributed system without introducing significant approximation errors
Performance Improvements for FDDI and CSMA/CD Protocols
The High-Performance Computing Initiative from the White House Office of Science and Technology Policy has defined 20 major challenges in science and engineering which are dependent on the solutions to a number of high-performance computing problems. One of the major areas of focus of this initiative is the development of gigabit rate networks to be used in environments such as the space station or a National Research and Educational Network (NREN).
The strategy here is to use existing network designs as building blocks for achieving higher rates, with the ultimate goal being a gigabit rate network. Two strategies which contribute to achieving this goal are examined in detail.1
FDDI2 is a token ring network based on fiber optics capable of a 100 Mbps rate. Both media access (MAC) and physical layer modifications are considered. A method is presented which allows one to determine maximum utilization based on the token-holding timer settings. Simulation results show that employing the second counter-rotating ring in combination with destination removal has a multiplicative effect greater than the effect which either of the factors have individually on performance. Two 100 Mbps rings can handle loads in the range of 400 to 500 Mbps for traffic with a uniform distribution and fixed packet size. Performance is dependent on the number of nodes, improving as the number increases. A wide range of environments are examined to illustrate robustness, and a method of implementation is discussed
Fog computing scheduling algorithm for smart city
With the advent of the number of smart devices across the globe, increasing the number of users using the Internet. The main aim of the fog computing (FC) paradigm is to connect huge number of smart objects (billions of object) that can make a bright future for smart cities. Due to the large deployments of smart devices, devices are expected to generate huge amounts of data and forward the data through the Internet. FC also refers to an edge computing framework that mitigates the issue by applying the process of knowledge discovery using a data analysis approach to the edges. Thus, the FC approaches can work together with the internet of things (IoT) world, which can build a sustainable infrastructure for smart cities. In this paper, we propose a scheduling algorithm namely the weighted round-robin (WRR) scheduling algorithm to execute the task from one fog node (FN) to another fog node to the cloud. Firstly, a fog simulator is used with the emergent concept of FC to design IoT infrastructure for smart cities. Then, spanning-tree routing (STP) protocol is used for data collection and routing. Further, 5G networks are proposed to establish fast transmission and communication between users. Finally, the performance of our proposed system is evaluated in terms of response time, latency, and amount of data used
- âŚ