47 research outputs found

    Mobile Ad Hoc Networks

    Get PDF
    Guiding readers through the basics of these rapidly emerging networks to more advanced concepts and future expectations, Mobile Ad hoc Networks: Current Status and Future Trends identifies and examines the most pressing research issues in Mobile Ad hoc Networks (MANETs). Containing the contributions of leading researchers, industry professionals, and academics, this forward-looking reference provides an authoritative perspective of the state of the art in MANETs. The book includes surveys of recent publications that investigate key areas of interest such as limited resources and the mobility of mobile nodes. It considers routing, multicast, energy, security, channel assignment, and ensuring quality of service. Also suitable as a text for graduate students, the book is organized into three sections: Fundamentals of MANET Modeling and Simulation—Describes how MANETs operate and perform through simulations and models Communication Protocols of MANETs—Presents cutting-edge research on key issues, including MAC layer issues and routing in high mobility Future Networks Inspired By MANETs—Tackles open research issues and emerging trends Illustrating the role MANETs are likely to play in future networks, this book supplies the foundation and insight you will need to make your own contributions to the field. It includes coverage of routing protocols, modeling and simulations tools, intelligent optimization techniques to multicriteria routing, security issues in FHAMIPv6, connecting moving smart objects to the Internet, underwater sensor networks, wireless mesh network architecture and protocols, adaptive routing provision using Bayesian inference, and adaptive flow control in transport layer using genetic algorithms

    Improving Multicast Communications Over Wireless Mesh Networks

    Get PDF
    In wireless mesh networks (WMNs) the traditional approach to shortest path tree based multicasting is to cater for the needs of the poorest performingnode i.e. the maximum permitted multicast line rate is limited to the lowest line rate used by the individual Child nodes on a branch. In general, this meansfixing the line rate to its minimum value and fixing the transmit power to its maximum permitted value. This simplistic approach of applying a single multicast rate for all nodes in the multicast group results in a sub-optimal trade-off between the mean network throughput and coverage area that does not allow for high bandwidth multimedia applications to be supported. By relaxing this constraint and allowing multiple line rates to be used, the mean network throughput can be improved. This thesis presents two methods that aim to increase the mean network throughput through the use of multiple line rates by the forwarding nodes. This is achieved by identifying the Child nodes responsible for reducing the multicast group rate. The first method identifies specific locations for the placement of relay nodes which allows for higher multicast branch line rates to be used. The second method uses a power control algorithm to tune the transmit power to allow for higher multicast branch line rates. The use of power control also helps to reduce the interference caused to neighbouring nodes.Through extensive computer simulation it can be shown that these two methods can lead to a four-fold gain in the mean network throughput undertypical WMN operating conditions compared with the single line rate case

    Game theoretic models for resource sharing in wireless networks

    Get PDF
    Wireless communications have been recently characterized by rapid proliferation of wireless networks, impressive growth of standard and technologies, evolution of the end-user terminals, and increasing demand in the wireless spectrum. New, more flexible schemes for the management of the available resources, from both the user and the network side, are necessary in order to improve the efficiency in the usage of the available resources.This work aims at shedding light on the performance modeling of radio resource sharing/allocation situations. Since, in general, the quality of service perceived by a system (e.g., user, network) strictly depends on the behavior of the other entities, and the involved interactions are mainly competitive, this work introduces a framework based on non–cooperative game theoretic tools. Furthermore, non–cooperative game theory is suitable in distributed networks, where control and management are inherently decentralized.First, we consider the case in which many users have to make decisions on which wireless access point to connect to. In this scenario, the quality perceived by the users mainly depends on the number of other users choosing the very same accessing opportunity. In this context, we also consider two–stage games where network make decisions on how to use the available resources, and users react to this selecting the network that maximizes their satisfaction. Then, we refer to the problem of spectrum sharing, where users directly compete for portions of the available spectrum. Finally, we provide a more complex model where the users utility function is based on the Shannon rate. The aim of this second part is to provide a better representation of the satisfaction perceived by the users, i.e., in terms of achievable throughput. Due to the complexity of the game model, we first provide a complete analytical analysis of the two–user case. Then, we extend the model to the N–user case. We mainly analyze this game through simulations. Finally, inspired by the results obtained numerically, we introduce stochastic geometry in the analysis of spectrum games in order to predict the performance of the game in large networks.Ph.D., Electrical Engineering -- Drexel University, 201

    Structure, Dynamics and Self-Organization in Recurrent Neural Networks: From Machine Learning to Theoretical Neuroscience

    Get PDF
    At a first glance, artificial neural networks, with engineered learning algorithms and carefully chosen nonlinearities, are nothing like the complicated self-organized spiking neural networks studied by theoretical neuroscientists. Yet, both adapt to their inputs, keep information from the past in their state space and are able of learning, implying that some information processing principles should be common to both. In this thesis we study those principles by incorporating notions of systems theory, statistical physics and graph theory into artificial neural networks and theoretical neuroscience models. % TO DO: What is different in this thesis? -> classical signal processing with complex systems on top The starting point for this thesis is \ac{RC}, a learning paradigm used both in machine learning\cite{jaeger2004harnessing} and in theoretical neuroscience\cite{maass2002real}. A neural network in \ac{RC} consists of two parts, a reservoir – a directed and weighted network of neurons that projects the input time series onto a high dimensional space – and a readout which is trained to read the state of the neurons in the reservoir and combine them linearly to give the desired output. In classical \ac{RC}, the reservoir is randomly initialized and left untrained, which alleviates the training costs in comparison to other recurrent neural networks. However, this lack of training implies that reservoirs are not adapted to specific tasks and thus their performance is often lower than that of other neural networks. Our contribution has been to show how knowledge about a task can be integrated into the reservoir architecture, so that reservoirs can be tailored to specific problems without training. We do this design by identifying two features that are useful for machine learning: the memory of the reservoir and its power spectra. First we show that the correlations between neurons limit the capacity of the reservoir to retain traces of previous inputs, and demonstrate that those correlations are controlled by moduli of the eigenvalues of the adjacency matrix of the reservoir. Second, we prove that when the reservoir resonates at the frequencies that are present on the desired output signal, the performance of the readout increases. Knowing the features of the reservoir dynamics that we need, the next question is how to impose them. The simplest way to design a network with that resonates at a certain frequency is by adding cycles, which act as feedback loops, but this also induces correlations and hence memory modifications. To disentangle the frequencies and the memory design, we studied how the addition of cycles modifies the eigenvalues in the adjacency matrix of the network. Surprisingly, the shape of the eigenvalues is quite beautiful \cite{aceituno2019universal} and can be characterized using random matrix theory tools. Combining this knowledge with our result relating eigenvalues and correlations, we designed an heuristic that tailors reservoirs to specific tasks and showed that it improves upon state of the art \ac{RC} in three different machine learning tasks. Although this idea works in the machine learning version of \ac{RC}, there is one fundamental problem when we try to translate to the world of theoretical neuroscience: the proposed frequency adaptation requires prior knowledge of the task, which might not be plausible in a biological neural network. Therefore the following questions are whether those resonances can emerge by unsupervised learning, and which kind of learning rules would be required. Remarkably, these resonances can be induced by the well-known Spike Time-Dependent Plasticity (STDP) combined with homeostatic mechanisms. We show this by deriving two self-consistent equations: one where the activity of every neuron can be calculated from its synaptic weights and its external inputs and a second one where the synaptic weights can be obtained from the neural activity. By considering spatio-temporal symmetries in our inputs we obtained two families of solutions to those equations where a periodic input is enhanced by the neural network after STDP. This approach shows that periodic and quasiperiodic inputs can induce resonances that agree with the aforementioned \ac{RC} theory. Those results, although rigorous, are expressed on a language of statistical physics and cannot be easily tested or verified in real, scarce data. To make them more accessible to the neuroscience community we showed that latency reduction, a well-known effect of STDP\cite{song2000competitive} which has been experimentally observed \cite{mehta2000experience}, generates neural codes that agree with the self-consistency equations and their solutions. In particular, this analysis shows that metabolic efficiency, synchronization and predictions can emerge from that same phenomena of latency reduction, thus closing the loop with our original machine learning problem. To summarize, this thesis exposes principles of learning recurrent neural networks that are consistent with adaptation in the nervous system and also improve current machine learning methods. This is done by leveraging features of the dynamics of recurrent neural networks such as resonances and correlations in machine learning problems, then imposing the required dynamics into reservoir computing through control theory notions such as feedback loops and spectral analysis. Then we assessed the plausibility of such adaptation in biological networks, deriving solutions from self-organizing processes that are biologically plausible and align with the machine learning prescriptions. Finally, we relate those processes to learning rules in biological neurons, showing how small local adaptations of the spike times can lead to neural codes that are efficient and can be interpreted in machine learning terms
    corecore