730 research outputs found
A measurement-based approach to service modeling and bandwidth estimation in IEEE 802.11 wireless networks
[no abstract
Understanding Fairness and its Impact on Quality of Service in IEEE 802.11
The Distributed Coordination Function (DCF) aims at fair and efficient medium
access in IEEE 802.11. In face of its success, it is remarkable that there is
little consensus on the actual degree of fairness achieved, particularly
bearing its impact on quality of service in mind. In this paper we provide an
accurate model for the fairness of the DCF. Given M greedy stations we assume
fairness if a tagged station contributes a share of 1/M to the overall number
of packets transmitted. We derive the probability distribution of fairness
deviations and support our analytical results by an extensive set of
measurements. We find a closed-form expression for the improvement of long-term
over short-term fairness. Regarding the random countdown values we quantify the
significance of their distribution whereas we discover that fairness is largely
insensitive to the distribution parameters. Based on our findings we view the
DCF as emulating an ideal fair queuing system to quantify the deviations from a
fair rate allocation. We deduce a stochastic service curve model for the DCF to
predict packet delays in IEEE 802.11. We show how a station can estimate its
fair bandwidth share from passive measurements of its traffic arrivals and
departures
Semantic and effective communications
Shannon and Weaver categorized communications into three levels of problems: the technical problem, which tries to answer the question "how accurately can the symbols of communication be transmitted?"; the semantic problem, which asks the question "how precisely do the transmitted symbols convey the desired meaning?"; the effectiveness problem, which strives to answer the question "how effectively does the received meaning affect conduct in the desired way?". Traditionally, communication technologies mainly addressed the technical problem, ignoring the semantics or the effectiveness problems.
Recently, there has been increasing interest to address the higher level semantic and effectiveness problems, with proposals ranging from semantic to goal oriented communications. In this thesis, we propose to formulate the semantic problem as a joint source-channel coding (JSCC) problem and the effectiveness problem as a multi-agent partially observable Markov decision process (MA-POMDP). As such, for the semantic problem, we propose DeepWiVe, the first-ever end-to-end JSCC video transmission scheme that leverages the power of deep neural networks (DNNs) to directly map video signals to channel symbols, combining video compression, channel coding, and modulation steps into a single neural transform. We also further show that it is possible to use predefined constellation designs as well as secure the physical layer communication against eavesdroppers for deep learning (DL) driven JSCC schemes, making such schemes much more viable for deployment in the real world.
For the effectiveness problem, we propose a novel formulation by considering multiple agents communicating over a noisy channel in order to achieve better coordination and cooperation in a multi-agent reinforcement learning (MARL) framework. Specifically, we consider a MA-POMDP, in which the agents, in addition to interacting with the environment, can also communicate with each other over a noisy communication channel. The noisy communication channel is considered explicitly as part of the dynamics of the environment, and the message each agent sends is part of the action that the agent can take. As a result, the agents learn not only to collaborate with each other but also to communicate "effectively'' over a noisy channel. Moreover, we show that this framework generalizes both the semantic and technical problems. In both instances, we show that the resultant communication scheme is superior to one where the communication is considered separately from the underlying semantic or goal of the problem.Open Acces
Prediction-based techniques for the optimization of mobile networks
MenciĂłn Internacional en el tĂtulo de doctorMobile cellular networks are complex system whose behavior is characterized by the superposition
of several random phenomena, most of which, related to human activities, such as mobility,
communications and network usage. However, when observed in their totality, the many individual
components merge into more deterministic patterns and trends start to be identifiable and
predictable.
In this thesis we analyze a recent branch of network optimization that is commonly referred to
as anticipatory networking and that entails the combination of prediction solutions and network
optimization schemes. The main intuition behind anticipatory networking is that knowing in
advance what is going on in the network can help understanding potentially severe problems and
mitigate their impact by applying solution when they are still in their initial states. Conversely,
network forecast might also indicate a future improvement in the overall network condition (i.e.
load reduction or better signal quality reported from users). In such a case, resources can be
assigned more sparingly requiring users to rely on buffered information while waiting for the
better condition when it will be more convenient to grant more resources.
In the beginning of this thesis we will survey the current anticipatory networking panorama
and the many prediction and optimization solutions proposed so far. In the main body of the work,
we will propose our novel solutions to the problem, the tools and methodologies we designed to
evaluate them and to perform a real world evaluation of our schemes.
By the end of this work it will be clear that not only is anticipatory networking a very promising
theoretical framework, but also that it is feasible and it can deliver substantial benefit to current
and next generation mobile networks. In fact, with both our theoretical and practical results we
show evidences that more than one third of the resources can be saved and even larger gain can
be achieved for data rate enhancements.Programa Oficial de Doctorado en IngenierĂa TelemĂĄticaPresidente: Albert Banchs Roca.- Presidente: Pablo Serrano Yañez-Mingot.- Secretario: Jorge OrtĂn Gracia.- Vocal: Guevara Noubi
Adaptive Prioritized Random Linear Coding and Scheduling for Layered Data Delivery From Multiple Servers
In this paper, we deal with the problem of jointly determining the optimal coding strategy and the scheduling decisions when receivers obtain layered data from multiple servers. The layered data is encoded by means of prioritized random linear coding (PRLC) in order to be resilient to channel loss while respecting the unequal levels of importance in the data, and data blocks are transmitted simultaneously in order to reduce decoding delays and improve the delivery performance. We formulate the optimal coding and scheduling decisions problem in our novel framework with the help of Markov decision processes (MDP), which are effective tools for modeling adapting streaming systems. Reinforcement learning approaches are then proposed to derive reduced computational complexity solutions to the adaptive coding and scheduling problems. The novel reinforcement learning approaches and the MDP solution are examined in an illustrative example for scalable video transmission . Our methods offer large performance gains over competing methods that deliver the data blocks sequentially. The experimental evaluation also shows that our novel algorithms offer continuous playback and guarantee small quality variations which is not the case for baseline solutions. Finally, our work highlights the advantages of reinforcement learning algorithms to forecast the temporal evolution of data demands and to decide the optimal coding and scheduling decisions
Packet level measurement over wireless access
PhDPerformance Measurement of the IP packet networks mainly comprise of monitoring the network performance in terms of packet losses and delays. If used appropriately, these network parameters (i.e. delay, loss and bandwidth etc) can indicate the performance status of the network and they can be used in fault and performance monitoring, network provisioning, and traffic engineering. Globally, there is a growing need for accurate network measurement to support the commercial use of IP networks. In wireless networks, transmission losses and communication delays strongly affect the performance of the network. Compared to wired networks, wireless networks experience higher levels of data dropouts, and corruption due to issues of channel fading, noise, interference and mobility. Performance monitoring is a vital element in the commercial future of broadband packet networking and the ability to guarantee quality of service in such networks is implicit in Service Level Agreements.
Active measurements are performed by injecting probes, and this is widely used to determine the end to end performance. End to end delay in wired networks has been extensively investigated, and in this thesis we report on the accuracy achieved by probing for end to end delay over a wireless scenario. We have compared two probing techniques i.e. Periodic and Poisson probing, and estimated the absolute error for both. The simulations have been performed for single hop and multi- hop wireless networks.
In addition to end to end latency, Active measurements have also been performed for packet loss rate. The simulation based analysis has been tried under different traffic scenarios using Poisson Traffic Models. We have sampled the user traffic using Periodic probing at different rates for single hop and multiple hop wireless scenarios.
5
Active probing becomes critical at higher values of load forcing the network to saturation much earlier. We have evaluated the impact of monitoring overheads on the user traffic, and show that even small amount of probing overhead in a wireless medium can cause large degradation in network performance. Although probing at high rate provides a good estimation of delay distribution of user traffic with large variance yet there is a critical tradeoff between the accuracy of measurement and the packet probing overhead. Our results suggest that active probing is highly affected by probe size, rate, pattern, traffic load, and nature of shared medium, available bandwidth and the burstiness of the traffic
Optimized Packet Scheduling in Multiview Video Navigation Systems
In multiview video systems, multiple cameras generally acquire the same scene
from different perspectives, such that users have the possibility to select
their preferred viewpoint. This results in large amounts of highly redundant
data, which needs to be properly handled during encoding and transmission over
resource-constrained channels. In this work, we study coding and transmission
strategies in multicamera systems, where correlated sources send data through a
bottleneck channel to a central server, which eventually transmits views to
different interactive users. We propose a dynamic correlation-aware packet
scheduling optimization under delay, bandwidth, and interactivity constraints.
The optimization relies both on a novel rate-distortion model, which captures
the importance of each view in the 3D scene reconstruction, and on an objective
function that optimizes resources based on a client navigation model. The
latter takes into account the distortion experienced by interactive clients as
well as the distortion variations that might be observed by clients during
multiview navigation. We solve the scheduling problem with a novel
trellis-based solution, which permits to formally decompose the multivariate
optimization problem thereby significantly reducing the computation complexity.
Simulation results show the gain of the proposed algorithm compared to baseline
scheduling policies. More in details, we show the gain offered by our dynamic
scheduling policy compared to static camera allocation strategies and to
schemes with constant coding strategies. Finally, we show that the best
scheduling policy consistently adapts to the most likely user navigation path
and that it minimizes distortion variations that can be very disturbing for
users in traditional navigation systems
Advanced flight control system study
The architecture, requirements, and system elements of an ultrareliable, advanced flight control system are described. The basic criteria are functional reliability of 10 to the minus 10 power/hour of flight and only 6 month scheduled maintenance. A distributed system architecture is described, including a multiplexed communication system, reliable bus controller, the use of skewed sensor arrays, and actuator interfaces. Test bed and flight evaluation program are proposed
What broke where for distributed and parallel applications â a whodunit story
Detection, diagnosis and mitigation of performance problems in today\u27s large-scale distributed and parallel systems is a difficult task. These large distributed and parallel systems are composed of various complex software and hardware components. When the system experiences some performance or correctness problem, developers struggle to understand the root cause of the problem and fix in a timely manner. In my thesis, I address these three components of the performance problems in computer systems. First, we focus on diagnosing performance problems in large-scale parallel applications running on supercomputers. We developed techniques to localize the performance problem for root-cause analysis. Parallel applications, most of which are complex scientific simulations running in supercomputers, can create up to millions of parallel tasks that run on different machines and communicate using the message passing paradigm. We developed a highly scalable and accurate automated debugging tool called PRODOMETER, which uses sophisticated algorithms to first, create a logical progress dependency graph of the tasks to highlight how the problem spread through the system manifesting as a system-wide performance issue. Second, uses this logical progress dependence graph to identify the task where the problem originated. Finally, PRODOMETER pinpoints the code region corresponding to the origin of the bug. Second, we developed a tool-chain that can detect performance anomaly using machine-learning techniques and can achieve very low false positive rate. Our input-aware performance anomaly detection system consists of a scalable data collection framework to collect performance related metrics from different granularity of code regions, an offline model creation and prediction-error characterization technique, and a threshold based anomaly-detection-engine for production runs. Our system requires few training runs and can handle unknown inputs and parameter combinations by dynamically calibrating the anomaly detection threshold according to the characteristics of the input data and the characteristics of the prediction-error of the models. Third, we developed performance problem mitigation scheme for erasure-coded distributed storage systems. Repair operations of the failed blocks in erasure-coded distributed storage system take really long time in networked constrained data-centers. The reason being, during the repair operation for erasure-coded distributed storage, a lot of data from multiple nodes are gathered into a single node and then a mathematical operation is performed to reconstruct the missing part. This process severely congests the links toward the destination where newly recreated data is to be hosted. We proposed a novel distributed repair technique, called Partial-Parallel-Repair (PPR) that performs this reconstruction in parallel on multiple nodes and eliminates network bottlenecks, and as a result, greatly speeds up the repair process. Fourth, we study how for a class of applications, performance can be improved (or performance problems can be mitigated) by selectively approximating some of the computations. For many applications, the main computation happens inside a loop that can be logically divided into a few temporal segments, we call phases. We found that while approximating the initial phases might severely degrade the quality of the results, approximating the computation for the later phases have very small impact on the final quality of the result. Based on this observation, we developed an optimization framework that for a given budget of quality-loss, would find the best approximation settings for each phase in the execution
Mathematical analysis of scheduling policies in peer-to-peer video streaming networks
Las redes de pares son comunidades virtuales autogestionadas, desarrolladas en la capa de aplicaciĂłn sobre la infraestructura de Internet, donde los usuarios (denominados pares) comparten recursos (ancho de banda, memoria, procesamiento) para alcanzar un fin comĂșn. La distribuciĂłn de video representa la aplicaciĂłn mĂĄs desafiante, dadas las limitaciones de ancho de banda. Existen bĂĄsicamente tres servicios de video. El mĂĄs simple es la descarga, donde un conjunto de servidores posee el contenido original, y los usuarios deben descargar completamente este contenido previo a su reproducciĂłn. Un segundo servicio se denomina video bajo demanda, donde los pares se unen a una red virtual siempre que inicien una solicitud de un contenido de video, e inician una descarga progresiva en lĂnea. El Ășltimo servicio es video en vivo, donde el contenido de video es generado, distribuido y visualizado simultĂĄneamente. En esta tesis se estudian aspectos de diseño para la distribuciĂłn de video en vivo y bajo demanda. Se presenta un anĂĄlisis matemĂĄtico de estabilidad y capacidad de arquitecturas de distribuciĂłn bajo demanda hĂbridas, asistidas por pares. Los pares inician descargas concurrentes de mĂșltiples contenidos, y se desconectan cuando lo desean. Se predice la evoluciĂłn esperada del sistema asumiendo proceso Poisson de arribos y egresos exponenciales, mediante un modelo determinĂstico de fluidos. Un sub-modelo de descargas secuenciales (no simultĂĄneas) es globalmente y estructuralmente estable, independientemente de los parĂĄmetros de la red. Mediante la Ley de Little se determina el tiempo medio de residencia de usuarios en un sistema bajo demanda secuencial estacionario. Se demuestra teĂłricamente que la filosofĂa hĂbrida de cooperaciĂłn entre pares siempre desempeña mejor que la tecnologĂa pura basada en cliente-servidor
- âŠ