35,047 research outputs found

    The Network Effects of Prefetching

    Full text link
    Prefetching has been shown to be an effective technique for reducing user perceived latency in distributed systems. In this paper we show that even when prefetching adds no extra traffic to the network, it can have serious negative performance effects. Straightforward approaches to prefetching increase the burstiness of individual sources, leading to increased average queue sizes in network switches. However, we also show that applications can avoid the undesirable queueing effects of prefetching. In fact, we show that applications employing prefetching can significantly improve network performance, to a level much better than that obtained without any prefetching at all. This is because prefetching offers increased opportunities for traffic shaping that are not available in the absence of prefetching. Using a simple transport rate control mechanism, a prefetching application can modify its behavior from a distinctly ON/OFF entity to one whose data transfer rate changes less abruptly, while still delivering all data in advance of the user's actual requests

    Analysis of a batch-service queue with variable service capacity, correlated customer types and generally distributed class-dependent service times

    Get PDF
    Queueing models with batch service have been studied frequently, for instance in the domain of telecommunications or manufacturing. Although the batch server's capacity may be variable in practice, only a few authors have included variable capacity in their models. We analyse a batch server with multiple customer classes and a variable service capacity that depends on both the number of waiting customers and their classes. The service times are generally distributed and class-dependent. These features complicate the analysis in a non-trivial way. We tackle it by examining the system state at embedded points, and studying the resulting Markov Chain. We first establish the joint probability generating function (pgf) of the service capacity and the number of customers left behind in the queue immediately after service initiation epochs. From this joint pgf, we extract the pgf for the number of customers in the queue and in the system respectively at service initiation epochs and departure epochs, and the pgf of the actual server capacity. Combined with additional techniques, we also obtain the pgf of the queue and system content at customer arrival epochs and random slot boundaries, and the pgf of the delay of a random customer. In the numerical experiments, we focus on the impact of correlation between the classes of consecutive customers, and on the influence of different service time distributions on the system performance. (C) 2019 Elsevier B.V. All rights reserved
    • …
    corecore