16,507 research outputs found

    An Efficient Medium Access Control Strategy for High Speed WDM Multiaccess Networks

    Get PDF
    A medium access control (MAC) strategy that accounts for the limited tunability of present-day lasers and filters and yet supports a large total number of wavelengths in the network is proposed. Full interconnectivity, contention-free access and a high value of concurrency are achieved by dividing the network into disjunct subnetworks on a wavelength basis and by reconfiguring these subnetworks on a time basis. Each subnetwork allows for simplified access to be implemented with fast tunable transceivers each assessing only a moderate number of wavelengths. A performance analysis shows that this concept is most efficient when applied to a high-level broadband interconnection metropolitan area network (MAN

    When Channel Bonding is Beneficial for Opportunistic Spectrum Access Networks

    Full text link
    Transmission over multiple frequency bands combined into one logical channel speeds up data transfer for wireless networks. On the other hand, the allocation of multiple channels to a single user decreases the probability of finding a free logical channel for new connections, which may result in a network-wide throughput loss. While this relationship has been studied experimentally, especially in the WLAN configuration, little is known on how to analytically model such phenomena. With the advent of Opportunistic Spectrum Access (OSA) networks, it is even more important to understand the circumstances in which it is beneficial to bond channels occupied by primary users with dynamic duty cycle patterns. In this paper we propose an analytical framework which allows the investigation of the average channel throughput at the medium access control layer for OSA networks with channel bonding enabled. We show that channel bonding is generally beneficial, though the extent of the benefits depend on the features of the OSA network, including OSA network size and the total number of channels available for bonding. In addition, we show that performance benefits can be realized by adaptively changing the number of bonded channels depending on network conditions. Finally, we evaluate channel bonding considering physical layer constraints, i.e. throughput reduction compared to the theoretical throughput of a single virtual channel due to a transmission power limit for any bonding size.Comment: accepted to IEEE Transactions on Wireless Communication

    CMD: A Multi-Channel Coordination Scheme for Emergency Message Dissemination in IEEE 1609.4

    Full text link
    In the IEEE 1609.4 legacy standard for multi-channel communications in vehicular ad hoc networks(VANETs), the control channel (CCH) is dedicated to broadcast safety messages while the service channels (SCH's) are dedicated to transmit infotainment service content. However, the SCH can be used as an alternative to transmit high priority safety messages in the event that they are invoked during the service channel interval (SCHI). This implies that there is a need to transmit safety messages across multiple available utilized channels to ensure that all vehicles receive the safety message. Transmission across multiple SCH's using the legacy IEEE 1609.4 requires multiple channel switching and therefore introduces further end-to-end delays. Given that safety messaging is a life critical application, it is important that optimal end-to-end delay performance is derived in multi-channel VANET scenarios to ensure reliable safety message dissemination. To tackle this challenge, three primary contributions are in this article: first, a channel coordinator selection approach based on the least average separation distance (LAD) to the vehicles that expect to tune to other SCH's and operates during the control channel interval (CCHI) is proposed. Second, a model to determine the optimal time intervals in which CMD operates during the CCHI is proposed. Third, a contention back-off mechanism for safety message transmission during the SCHI is proposed. Computer simulations and mathematical analysis show that CMD performs better than the legacy IEEE 1609.4 and a selected state-of-the-art multi-channel message dissemination schemes in terms of end-to-end delay and packet reception ratio.Comment: 15 pages, 10 figures, 7 table

    Distributed Game Theoretic Optimization and Management of Multichannel ALOHA Networks

    Full text link
    The problem of distributed rate maximization in multi-channel ALOHA networks is considered. First, we study the problem of constrained distributed rate maximization, where user rates are subject to total transmission probability constraints. We propose a best-response algorithm, where each user updates its strategy to increase its rate according to the channel state information and the current channel utilization. We prove the convergence of the algorithm to a Nash equilibrium in both homogeneous and heterogeneous networks using the theory of potential games. The performance of the best-response dynamic is analyzed and compared to a simple transmission scheme, where users transmit over the channel with the highest collision-free utility. Then, we consider the case where users are not restricted by transmission probability constraints. Distributed rate maximization under uncertainty is considered to achieve both efficiency and fairness among users. We propose a distributed scheme where users adjust their transmission probability to maximize their rates according to the current network state, while maintaining the desired load on the channels. We show that our approach plays an important role in achieving the Nash bargaining solution among users. Sequential and parallel algorithms are proposed to achieve the target solution in a distributed manner. The efficiencies of the algorithms are demonstrated through both theoretical and simulation results.Comment: 34 pages, 6 figures, accepted for publication in the IEEE/ACM Transactions on Networking, part of this work was presented at IEEE CAMSAP 201

    Spatial Performance Analysis and Design Principles for Wireless Peer Discovery

    Full text link
    In wireless peer-to-peer networks that serve various proximity-based applications, peer discovery is the key to identifying other peers with which a peer can communicate and an understanding of its performance is fundamental to the design of an efficient discovery operation. This paper analyzes the performance of wireless peer discovery through comprehensively considering the wireless channel, spatial distribution of peers, and discovery operation parameters. The average numbers of successfully discovered peers are expressed in closed forms for two widely used channel models, i.e., the interference limited Nakagami-m fading model and the Rayleigh fading model with nonzero noise, when peers are spatially distributed according to a homogeneous Poisson point process. These insightful expressions lead to the design principles for the key operation parameters including the transmission probability, required amount of wireless resources, level of modulation and coding scheme (MCS), and transmit power. Furthermore, the impact of shadowing on the spatial performance and suggested design principles is evaluated using mathematical analysis and simulations.Comment: 12 pages (double columns), 10 figures, 1 table, to appear in the IEEE Transactions on Wireless Communication

    Compressive Source Separation: Theory and Methods for Hyperspectral Imaging

    Get PDF
    With the development of numbers of high resolution data acquisition systems and the global requirement to lower the energy consumption, the development of efficient sensing techniques becomes critical. Recently, Compressed Sampling (CS) techniques, which exploit the sparsity of signals, have allowed to reconstruct signal and images with less measurements than the traditional Nyquist sensing approach. However, multichannel signals like Hyperspectral images (HSI) have additional structures, like inter-channel correlations, that are not taken into account in the classical CS scheme. In this paper we exploit the linear mixture of sources model, that is the assumption that the multichannel signal is composed of a linear combination of sources, each of them having its own spectral signature, and propose new sampling schemes exploiting this model to considerably decrease the number of measurements needed for the acquisition and source separation. Moreover, we give theoretical lower bounds on the number of measurements required to perform reconstruction of both the multichannel signal and its sources. We also proposed optimization algorithms and extensive experimentation on our target application which is HSI, and show that our approach recovers HSI with far less measurements and computational effort than traditional CS approaches.Comment: 32 page
    corecore