4 research outputs found

    A systematic approach to the design of complex systems : application to DBMS design and evaluation

    Get PDF
    Bibliography: leaves 80-85.R.C. Andreu, S.E. Madnick

    Modeling of Physical Database Design and Performance Analysis With Emphasis on VSAM Files.

    Get PDF
    Growth in the size of a database is reflected in deterioration of database performance. Since deterioration is related to the structure of the file, the performance efficiency involves the design of a physical database and the proper management of it. This research addresses a modeling procedure of a physical database design considering both records insertions and deletions. The model describes the behavior of a physical database in a VSAM file environment, and is extended to the issue of database reorganization through a cost analysis. The cost of accessing the database increases due to the physical disorganization of the database caused by records updates and insertions. A cost function that describes this excess cost is defined. As a remedy of the performance deterioration, database reorganization is required. Optimum reorganization points are obtained as a tradeoff between the excessive costs and the reorganization cost. Numerical examples based on the characteristics of IBM 3380 are given

    Attribute-Level Versioning: A Relational Mechanism for Version Storage and Retrieval

    Get PDF
    Data analysts today have at their disposal a seemingly endless supply of data and repositories hence, datasets from which to draw. New datasets become available daily thus making the choice of which dataset to use difficult. Furthermore, traditional data analysis has been conducted using structured data repositories such as relational database management systems (RDBMS). These systems, by their nature and design, prohibit duplication for indexed collections forcing analysts to choose one value for each of the available attributes for an item in the collection. Often analysts discover two or more datasets with information about the same entity. When combining this data and transforming it into a form that is usable in an RDBMS, analysts are forced to deconflict the collisions and choose a single value for each duplicated attribute containing differing values. This deconfliction is the source of a considerable amount of guesswork and speculation on the part of the analyst in the absence of professional intuition. One must consider what is lost by discarding those alternative values. Are there relationships between the conflicting datasets that have meaning? Is each dataset presenting a different and valid view of the entity or are the alternate values erroneous? If so, which values are erroneous? Is there a historical significance of the variances? The analysis of modern datasets requires the use of specialized algorithms and storage and retrieval mechanisms to identify, deconflict, and assimilate variances of attributes for each entity encountered. These variances, or versions of attribute values, contribute meaning to the evolution and analysis of the entity and its relationship to other entities. A new, distinct storage and retrieval mechanism will enable analysts to efficiently store, analyze, and retrieve the attribute versions without unnecessary complexity or additional alterations of the original or derived dataset schemas. This paper presents technologies and innovations that assist data analysts in discovering meaning within their data and preserving all of the original data for every entity in the RDBMS
    corecore