13 research outputs found

    Minimizing stall time in single and parallel disk systems

    No full text
    We study integrated prefetching and caching problems following the work of Cao et al. and Kimbrel and Karlin. Cao et al. and Kimbrel and Karlin gave approximation algorithms for minimizing the total elapsed time in single and parallel disk settings. The total elapsed time is the sum of the processor stall times and the length of the request sequence to be served. We show that an optimum prefetching/caching schedule for a single disk problem can be computed in polynomial time, thereby settling an open question by Kimbrel and Karlin. For the parallel disk problem we give an approximation algorithm for minimizing stall time. Stall time is a more realistic and harder to approximate measure for this problem. All of our algorithms are based on a new approach which involves formulating the prefetching/caching problems as integer programs

    Out-of-core visualization using iterator-aware multidimensional prefetching

    Full text link

    COMPACT FORMULATION OF MULTICOMMODITY NETWORK FLOWS WITH APPLICATIONS TO THE BACKHAUL PROFIT MAXIMIZATION PROBLEM AND FIXED CHARGE NETWORK FLOW PROBLEM

    Get PDF
    The triples formulation is a compact formulation of multicommodity network flow problems that provides a different representation of flow than the traditional and widely used node-arc and arc-path approaches. In the literature, the triples formulation has been applied successfully to the maximum concurrent flow problem and to a network optimization problem with piecewise linear convex costs. This dissertation applies the triples formulation to the backhaul profit maximization problem (BPMP) and the fixed charge network flow problem (FCNF). It is shown that the triples representation of multicommodity flow significantly reduces the number of variables and constraints in the mixed integer programming formulations of the BPMP and FCNF. For the BPMP, this results in significantly faster solution times. For dense problem instances, the triples-based formulation of FCNF is found to produce better solutions than the node-arc formulation early in the branch-and-bound process. This observation leads to an effective hybrid method which combines the respective advantages of the smaller size of the triples formulation and the stronger linear programming relaxation of the node-arc formulation. In addition to empirical studies, the dissertation presents new theoretical results supporting the equivalence of the triples formulation to the node-arc and arc-path formulations. The dissertation also proposes a multi-criteria Composite Index Method (CIM) to compare the performance of alternative integer programming formulations of an optimization problem. Using the CIM, the decision maker assigns weights to problem instance sizes and multiple performance measures based on their relative importance for the given application. The weighting scheme is used to produce a single number that measures the relative improvement of one alternative over the other and provides a method to select the most effective approach when neither one dominates the other when tested on different sizes of problem instances. The dissertation demonstrates a successful application of the CIM to evaluate a series of eleven techniques for improving the node-arc and triples formulations of the BPMP previously proposed in the literature

    Integrated prefetching and caching in single and parallel disk systems

    Get PDF

    Prefetching techniques for client server object-oriented database systems

    Get PDF
    The performance of many object-oriented database applications suffers from the page fetch latency which is determined by the expense of disk access. In this work we suggest several prefetching techniques to avoid, or at least to reduce, page fetch latency. In practice no prediction technique is perfect and no prefetching technique can entirely eliminate delay due to page fetch latency. Therefore we are interested in the trade-off between the level of accuracy required for obtaining good results in terms of elapsed time reduction and the processing overhead needed to achieve this level of accuracy. If prefetching accuracy is high then the total elapsed time of an application can be reduced significantly otherwise if the prefetching accuracy is low, many incorrect pages are prefetched and the extra load on the client, network, server and disks decreases the whole system performance. Access pattern of object-oriented databases are often complex and usually hard to predict accurately. The ..

    Granite: A scientific database model and implementation

    Get PDF
    The principal goal of this research was to develop a formal comprehensive model for representing highly complex scientific data. An effective model should provide a conceptually uniform way to represent data and it should serve as a framework for the implementation of an efficient and easy-to-use software environment that implements the model. The dissertation work presented here describes such a model and its contributions to the field of scientific databases. In particular, the Granite model encompasses a wide variety of datatypes used across many disciplines of science and engineering today. It is unique in that it defines dataset geometry and topology as separate conceptual components of a scientific dataset. We provide a novel classification of geometries and topologies that has important practical implications for a scientific database implementation. The Granite model also offers integrated support for multiresolution and adaptive resolution data. Many of these ideas have been addressed by others, but no one has tried to bring them all together in a single comprehensive model. The datasource portion of the Granite model offers several further contributions. In addition to providing a convenient conceptual view of rectilinear data, it also supports multisource data. Data can be taken from various sources and combined into a unified view. The rod storage model is an abstraction for file storage that has proven an effective platform upon which to develop efficient access to storage. Our spatial prefetching technique is built upon the rod storage model, and demonstrates very significant improvement in access to scientific datasets, and also allows machines to access data that is far too large to fit in main memory. These improvements bring the extremely large datasets now being generated in many scientific fields into the realm of tractability for the ordinary researcher. We validated the feasibility and viability of the model by implementing a significant portion of it in the Granite system. Extensive performance evaluations of the implementation indicate that the features of the model can be provided in a user-friendly manner with an efficiency that is competitive with more ad hoc systems and more specialized application specific solutions

    Minimizing Stall Time in Single and Parallel Disk Systems

    No full text
    We study integrated prefetching and caching problems following the work of Cao et. al. [3] and Kimbrel and Karlin [14]. Cao et. al. and Kimbrel and Karlin gave approximation algorithms for minimizing the total elapsed time in single and parallel disk settings. The total elapsed time is the sum of the processor stall times and the length of the request sequence to be served. We show that an optimum prefetching/caching schedule for a single disk problem can be computed in polynomial time, thereby settling an open question by Kimbrel and Karlin. For the parallel disk problem we give an approximation algorithm for minimizing stall time. Stall time is an important and harder to approximate measure for this problem. All of our algorithms are based on a new approach which involves formulating the prefetching/caching problems as integer programs

    Minimizing Stall Time in Single and Parallel Disk Systems.

    No full text
    We study integrated prefetching and caching problems following the work of Cao et al. [1995] and Kimbrel and Karlin [1996]. Cao et al. and Kimbrel and Karlin gave approximation algorithms for minimizing the total elapsed time in single and parallel disk settings. The total elapsed time is the sum of the processor stall times and the length of the request sequence to be served. We show that an optimum prefetching/caching schedule for a single disk problem can be computed in polynomial time, thereby settling an open question by Kimbrel and Karlin. For the parallel disk problem, we give an approximation algorithm for minimizing stall time. The solution uses a few extra memory blocks in cache. Stall time is an important and harder to approximate measure for this problem. All of our algorithms are based on a new approach which involves formulating the prefetching/caching problems as linear programs
    corecore