6,013 research outputs found
Performance of active multicast congestion control
This paper aims to provide insight into the behavior of congestion control mechanisms for reliable multicast protocols. A multicast congestion control based on active networks has been proposed and simulated using ns-2 over a network topology obtained using the Tiers tool. The congestion control mechanism has been simulated under different network conditions and with different settings of its configuration parameters. The objective is to analyze its performance and the impact of the different configuration parameters on its behavior. The simulation results show that the performance of the protocol is good in terms of delay and bandwidth utilization. The compatibility of the protocol with TCP flows has not been demonstrated, but the simulations performed show that by altering the parameter settings, the proportion of total bandwidth taken up by the two types of flow, multicast and TCP, may be modified.Publicad
Smooth Multirate Multicast Congestion Control
A significant impediment to deployment of multicast services is the daunting technical complexity of developing, testing and validating congestion control protocols ïŹt for wide-area deployment. Protocols such as pgmcc and TFMCC have recently made considerable progress on the single rate case, i.e. where one dynamic reception rate is maintained for all receivers in the session. However, these protocols have limited applicability, since scaling to session sizes beyond tens of participants necessitates the use of multiple rate protocols. Unfortunately, while existing multiple rate protocols exhibit better scalability, they are both less mature than single rate protocols and suffer from high complexity.
We propose a new approach to multiple rate congestion control that leverages proven single rate congestion control methods by orchestrating an ensemble of independently controlled single rate sessions. We describe SMCC, a new multiple rate equation-based congestion control algorithm for layered multicast sessions that employs TFMCC as the primary underlying control mechanism for each layer. SMCC combines the benefits of TFMCC (smooth rate control, equation-based TCP friendliness) with the scalability and flexibility of multiple rates to provide a sound multiple rate multicast congestion control policy.National Science Foundation (ANI-9986397, ANI-0092196
Adaptive Multicast of Multi-Layered Video: Rate-Based and Credit-Based Approaches
Network architectures that can efficiently transport high quality, multicast
video are rapidly becoming a basic requirement of emerging multimedia
applications. The main problem complicating multicast video transport is
variation in network bandwidth constraints. An attractive solution to this
problem is to use an adaptive, multi-layered video encoding mechanism. In this
paper, we consider two such mechanisms for the support of video multicast; one
is a rate-based mechanism that relies on explicit rate congestion feedback from
the network, and the other is a credit-based mechanism that relies on
hop-by-hop congestion feedback. The responsiveness, bandwidth utilization,
scalability and fairness of the two mechanisms are evaluated through
simulations. Results suggest that while the two mechanisms exhibit performance
trade-offs, both are capable of providing a high quality video service in the
presence of varying bandwidth constraints.Comment: 11 page
Recommended from our members
Effective video multicast over wireless internet
With the rapid growth of wireless networks and great success of Internet video, wireless video services are expected to be widely deployed in the near future. As different types of wireless networks are converging into all IP networks, i.e., the Internet, it is important to study video delivery over the wireless Internet. This paper proposes a novel end-system based adaptation protocol calledWireless Hybrid Adaptation Layered Multicast (WHALM) protocol for layered video multicast over wireless Internet. In WHALM the sender dynamically collects bandwidth distribution from the receivers and uses an optimal layer rate allocation mechanism to reduce the mismatches between the coarse-grained layer subscription levels and the heterogeneous and dynamic rate requirements from the receivers, thus maximizing the degree of satisfaction of all the receivers in a multicast session. Based on sampling theory and theory of probability, we reduce the required number of bandwidth feedbacks to a reasonable degree and use a scalable feedback mechanism to control the feedback process practically. WHALM is also tuned to perform well in wireless networks by integrating an end-to-end loss differentiation algorithm (LDA) to differentiate error losses from congestion losses at the receiver side. With a series of simulation experiments over NS platform, WHALM has been proved to be able to greatly improve the degree of satisfaction of all the receivers while avoiding congestion collapse on the wireless Internet
QuickCast: Fast and Efficient Inter-Datacenter Transfers using Forwarding Tree Cohorts
Large inter-datacenter transfers are crucial for cloud service efficiency and
are increasingly used by organizations that have dedicated wide area networks
between datacenters. A recent work uses multicast forwarding trees to reduce
the bandwidth needs and improve completion times of point-to-multipoint
transfers. Using a single forwarding tree per transfer, however, leads to poor
performance because the slowest receiver dictates the completion time for all
receivers. Using multiple forwarding trees per transfer alleviates this
concern--the average receiver could finish early; however, if done naively,
bandwidth usage would also increase and it is apriori unclear how best to
partition receivers, how to construct the multiple trees and how to determine
the rate and schedule of flows on these trees. This paper presents QuickCast, a
first solution to these problems. Using simulations on real-world network
topologies, we see that QuickCast can speed up the average receiver's
completion time by as much as while only using more
bandwidth; further, the completion time for all receivers also improves by as
much as faster at high loads.Comment: [Extended Version] Accepted for presentation in IEEE INFOCOM 2018,
Honolulu, H
Maximum Production Of Transmission Messages Rate For Service Discovery Protocols
Minimizing the number of dropped User Datagram Protocol (UDP) messages in a network is regarded as a challenge by researchers. This issue represents serious problems for many protocols particularly those that depend on sending messages as part of their strategy, such us service discovery protocols. This paper proposes and evaluates an algorithm to predict the minimum period of time required between two or more consecutive messages and suggests the minimum queue sizes for the routers, to manage the traffic and minimise the number of dropped messages that has been caused by either congestion or queue overflow or both together. The algorithm has been applied to the Universal Plug and Play (UPnP) protocol using ns2 simulator. It was tested when the routers were connected in two configurations; as a centralized and de centralized. The message length and bandwidth of the links among the routers were taken in the consideration. The result shows Better improvement in number of dropped messages `among the routers
Analysis domain model for shared virtual environments
The field of shared virtual environments, which also
encompasses online games and social 3D environments, has a
system landscape consisting of multiple solutions that share great functional overlap. However, there is little system interoperability between the different solutions. A shared virtual environment has an associated problem domain that is highly complex raising difficult challenges to the development process, starting with the architectural design of the underlying system. This paper has two main contributions. The first contribution is a broad domain analysis of shared virtual environments, which enables developers to have a better understanding of the whole rather than the part(s). The second contribution is a reference domain model for discussing and describing solutions - the Analysis Domain Model
Poor Man's Content Centric Networking (with TCP)
A number of different architectures have been proposed in support of data-oriented or information-centric networking. Besides a similar visions, they share the need for designing a new networking architecture. We present an incrementally deployable approach to content-centric networking based upon TCP. Content-aware senders cooperate with probabilistically operating routers for scalable content delivery (to unmodified clients), effectively supporting opportunistic caching for time-shifted access as well as de-facto synchronous multicast delivery. Our approach is application protocol-independent and provides support beyond HTTP caching or managed CDNs. We present our protocol design along with a Linux-based implementation and some initial feasibility checks
- âŠ