587 research outputs found

    Model-driven dual caching For nomadic service-oriented architecture clients

    Get PDF
    Mobile devices have evolved over the years from resource constrained devices that supported only the most basic tasks to powerful handheld computing devices. However, the most significant step in the evolution of mobile devices was the introduction of wireless connectivity which enabled them to host applications that require internet connectivity such as email, web browsers and maybe most importantly smart/rich clients. Being able to host smart clients allows the users of mobile devices to seamlessly access the Information Technology (IT) resources of their organizations. One increasingly popular way of enabling access to IT resources is by using Web Services (WS). This trend has been aided by the rapid availability of WS packages/tools, most notably the efforts of the Apache group and Integrated Development Environment (IDE) vendors. But the widespread use of WS raises questions for users of mobile devices such as laptops or PDAs; how and if they can participate in WS. Unlike their “wired” counterparts (desktop computers and servers) they rely on a wireless network that is characterized by low bandwidth and unreliable connectivity.The aim of this thesis is to enable mobile devices to host Web Services consumers. It introduces a Model-Driven Dual Caching (MDDC) approach to overcome problems arising from temporarily loss of connectivity and fluctuations in bandwidth

    Effect of speculative prefetching on network load in distributed systems

    Full text link
    Previous studies in speculative prefetching focus on building and evaluating access models for the purpose of access prediction. This paper on the other hand investigates the performance of speculative prefetching. When prefetching is performed speculatively, there is bound to be an increase in the network load. Furthermore, the prefetched items must compete for space with existing cache occupants. These two factors-increased load and eviction of potentially useful cache entries-are considered in the analysis. We obtain the following conclusion: to maximise the improvement in access time, prefetch exclusively all items with access probabilities exceeding a certain threshold.<br /

    Web Caching and Prefetching with Cyclic Model Analysis of Web Object Sequences

    Get PDF
    Web caching is the process in which web objects are temporarily stored to reduce bandwidth consumption, server load and latency. Web prefetching is the process of fetching web objects from the server before they are actually requested by the client. Integration of caching and prefetching can be very beneficial as the two techniques can support each other. By implementing this integrated scheme in a client-side proxy, the perceived latency can be reduced for not one but many users. In this paper, we propose a new integrated caching and prefetching policy called the WCP-CMA which makes use of a profit-driven caching policy that takes into account the periodicity and cyclic behaviour of the web access sequences for deriving prefetching rules. Our experimental results have shown a 10%-15% increase in the hit ratios of the cached objects and 5%-10% decrease in delay compared to the existing schem

    Multi-access Coded Caching with Decentralized Prefetching

    Full text link
    An extension of coded caching referred to as multi-access coded caching where each user can access multiple caches and each cache can serve multiple users is considered in this paper. Most of the literature in multi-access coded caching focuses on cyclic wrap-around cache access where each user is allowed to access an exclusive set of consecutive caches only. In this paper, a more general framework of multi-access caching problem is considered in which each user is allowed to randomly connect to a specific number of caches and multiple users can access the same set of caches. For the proposed system model considering decentralized prefetching, a new delivery scheme is proposed and an expression for per user delivery rate is obtained. A lower bound on the delivery rate is derived using techniques from index coding. The proposed scheme is shown to be optimal among all the linear schemes under certain conditions. An improved delivery rate and a lower bound for the decentralized multi-access coded caching scheme with cyclic wrap-around cache access can be obtained as a special case. By giving specific values to certain parameters, the results of decentralized shared caching scheme and of conventional decentralized caching scheme can be recovered.Comment: 26 pages, 6 figures, 6 tables, Submitted to IEEE Transactions on Communication
    • …
    corecore