232 research outputs found

    Database machines in support of very large databases

    Get PDF
    Software database management systems were developed in response to the needs of early data processing applications. Database machine research developed as a result of certain performance deficiencies of these software systems. This thesis discusses the history of database machines designed to improve the performance of database processing and focuses primarily on the Teradata DBC/1012, the only successfully marketed database machine that supports very large databases today. Also reviewed is the response of IBM to the performance needs of its database customers; this response has been in terms of improvements in both software and hardware support for database processing. In conclusion, an analysis is made of the future of database machines, in particular the DBC/1012, in light of recent IBM enhancements and its immense customer base

    The NASA computer science research program plan

    Get PDF
    A taxonomy of computer science is included, one state of the art of each of the major computer science categories is summarized. A functional breakdown of NASA programs under Aeronautics R and D, space R and T, and institutional support is also included. These areas were assessed against the computer science categories. Concurrent processing, highly reliable computing, and information management are identified

    Design and Performance analysis of a relational replicated database systems

    Get PDF
    The hardware organization and software structure of a new database system are presented. This system, the relational replicated database system (RRDS), is based on a set of replicated processors operating on a partitioned database. Performance improvements and capacity growth can be obtained by adding more processors to the configuration. Based on designing goals a set of hardware and software design questions were developed. The system then evolved according to a five-phase process, based on simulation and analysis, which addressed and resolved the design questions. Strategies and algorithms were developed for data access, data placement, and directory management for the hardware organization. A predictive performance analysis was conducted to determine the extent to which original design goals were satisfied. The predictive performance results, along with an analytical comparison with three other relational multi-backend systems, provided information about the strengths and weaknesses of our design as well as a basis for future research

    Ontology based data warehousing for mining of heterogeneous and multidimensional data sources

    Get PDF
    Heterogeneous and multidimensional big-data sources are virtually prevalent in all business environments. System and data analysts are unable to fast-track and access big-data sources. A robust and versatile data warehousing system is developed, integrating domain ontologies from multidimensional data sources. For example, petroleum digital ecosystems and digital oil field solutions, derived from big-data petroleum (information) systems, are in increasing demand in multibillion dollar resource businesses worldwide. This work is recognized by Industrial Electronic Society of IEEE and appeared in more than 50 international conference proceedings and journals

    Data storage hierarchy systems for data base computers.

    Get PDF
    Thesis. 1979. Ph.D.--Massachusetts Institute of Technology. Alfred P. Sloan School of Management.MICROFICHE COPY AVAILABLE IN ARCHIVES AND DEWEY.Vita.Bibliography: p. 241-248.Ph.D

    Space station data system analysis/architecture study. Task 2: Options development DR-5. Volume 1: Technology options

    Get PDF
    The second task in the Space Station Data System (SSDS) Analysis/Architecture Study is the development of an information base that will support the conduct of trade studies and provide sufficient data to make key design/programmatic decisions. This volume identifies the preferred options in the technology category and characterizes these options with respect to performance attributes, constraints, cost, and risk. The technology category includes advanced materials, processes, and techniques that can be used to enhance the implementation of SSDS design structures. The specific areas discussed are mass storage, including space and round on-line storage and off-line storage; man/machine interface; data processing hardware, including flight computers and advanced/fault tolerant computer architectures; and software, including data compression algorithms, on-board high level languages, and software tools. Also discussed are artificial intelligence applications and hard-wire communications

    Data bases and data base systems related to NASA's aerospace program. A bibliography with indexes

    Get PDF
    This bibliography lists 1778 reports, articles, and other documents introduced into the NASA scientific and technical information system, 1975 through 1980

    Leveraging Non-Volatile Memory in Modern Storage Management Architectures

    Get PDF
    Non-volatile memory technologies (NVM) introduce a novel class of devices that combine characteristics of both storage and main memory. Like storage, NVM is not only persistent, but also denser and cheaper than DRAM. Like DRAM, NVM is byte-addressable and has lower access latency. In recent years, NVM has gained a lot of attention both in academia and in the data management industry, with views ranging from skepticism to over excitement. Some critics claim that NVM is not cheap enough to replace flash-based SSDs nor is it fast enough to replace DRAM, while others see it simply as a storage device. Supporters of NVM have observed that its low latency and byte-addressability requires radical changes and a complete rewrite of storage management architectures. This thesis takes a moderate stance between these two views. We consider that, while NVM might not replace flash-based SSD or DRAM in the near future, it has the potential to reduce the gap between them. Furthermore, treating NVM as a regular storage media does not fully leverage its byte-addressability and low latency. On the other hand, completely redesigning systems to be NVM-centric is impractical. Proposals that attempt to leverage NVM to simplify storage management result in completely new architectures that face the same challenges that are already well-understood and addressed by the traditional architectures. Therefore, we take three common storage management architectures as a starting point, and propose incremental changes to enable them to better leverage NVM. First, in the context of log-structured merge-trees, we investigate the impact of storing data in NVM, and devise methods to enable small granularity accesses and NVM-aware caching policies. Second, in the context of B+Trees, we propose to extend the buffer pool and describe a technique based on the concept of optimistic consistency to handle corrupted pages in NVM. Third, we employ NVM to enable larger capacity and reduced costs in a index+log key-value store, and combine it with other techniques to build a system that achieves low tail latency. This thesis aims to describe and evaluate these techniques in order to enable storage management architectures to leverage NVM and achieve increased performance and lower costs, without major architectural changes.:1 Introduction 1.1 Non-Volatile Memory 1.2 Challenges 1.3 Non-Volatile Memory & Database Systems 1.4 Contributions and Outline 2 Background 2.1 Non-Volatile Memory 2.1.1 Types of NVM 2.1.2 Access Modes 2.1.3 Byte-addressability and Persistency 2.1.4 Performance 2.2 Related Work 2.3 Case Study: Persistent Tree Structures 2.3.1 Persistent Trees 2.3.2 Evaluation 3 Log-Structured Merge-Trees 3.1 LSM and NVM 3.2 LSM Architecture 3.2.1 LevelDB 3.3 Persistent Memory Environment 3.4 2Q Cache Policy for NVM 3.5 Evaluation 3.5.1 Write Performance 3.5.2 Read Performance 3.5.3 Mixed Workloads 3.6 Additional Case Study: RocksDB 3.6.1 Evaluation 4 B+Trees 4.1 B+Tree and NVM 4.1.1 Category #1: Buffer Extension 4.1.2 Category #2: DRAM Buffered Access 4.1.3 Category #3: Persistent Trees 4.2 Persistent Buffer Pool with Optimistic Consistency 4.2.1 Architecture and Assumptions 4.2.2 Embracing Corruption 4.3 Detecting Corruption 4.3.1 Embracing Corruption 4.4 Repairing Corruptions 4.5 Performance Evaluation and Expectations 4.5.1 Checksums Overhead 4.5.2 Runtime and Recovery 4.6 Discussion 5 Index+Log Key-Value Stores 5.1 The Case for Tail Latency 5.2 Goals and Overview 5.3 Execution Model 5.3.1 Reactive Systems and Actor Model 5.3.2 Message-Passing Communication 5.3.3 Cooperative Multitasking 5.4 Log-Structured Storage 5.5 Networking 5.6 Implementation Details 5.6.1 NVM Allocation on RStore 5.6.2 Log-Structured Storage and Indexing 5.6.3 Garbage Collection 5.6.4 Logging and Recovery 5.7 Systems Operations 5.8 Evaluation 5.8.1 Methodology 5.8.2 Environment 5.8.3 Other Systems 5.8.4 Throughput Scalability 5.8.5 Tail Latency 5.8.6 Scans 5.8.7 Memory Consumption 5.9 Related Work 6 Conclusion Bibliography A PiBenc

    Information resources management, 1984-1989: A bibliography with indexes

    Get PDF
    This bibliography contains 768 annotated references to reports and journal articles entered into the NASA scientific and technical information database 1984 to 1989

    Métodos computacionais para otimização de desempenho em redes de imagem médica

    Get PDF
    Over the last few years, the medical imaging has consolidated its position as a major mean of clinical diagnosis. The amount of data generated by the medical imaging practice is increasing tremendously. As a result, repositories are turning into rich databanks of semi-structured data related to patients, ailments, equipment and other stakeholders involved in the medical imaging panorama. The exploration of these repositories for secondary uses of data promises to elevate the quality standards and efficiency of the medical practice. However, supporting these advanced usage scenarios in traditional institutional systems raises many technical challenges that are yet to be overcome. Moreover, the reported poor performance of standard protocols opened doors to the general usage of proprietary solutions, compromising the interoperability necessary for supporting these advanced scenarios. This thesis has researched, developed, and now proposes a series of computer methods and architectures intended to maximize the performance of multi-institutional medical imaging environments. The methods are intended to improve the performance of standard protocols for medical imaging content discovery and retrieval. The main goal is to use them to increase the acceptance of vendor-neutral solutions through the improvement of their performance. Moreover, it intends to promote the adoption of such standard technologies in advanced scenarios that are still a mirage nowadays, such as clinical research or data analytics directly on top of live institutional repositories. Finally, these achievements will facilitate the cooperation between healthcare institutions and researchers, resulting in an increment of healthcare quality and institutional efficiency.As diversas modalidades de imagem médica têm vindo a consolidar a sua posição dominante como meio complementar de diagnóstico. O número de procedimentos realizados e o volume de dados gerados aumentou significativamente nos últimos anos, colocando pressão nas redes e sistemas que permitem o arquivo e distribuição destes estudos. Os repositórios de estudos imagiológicos são fontes de dados ricas contendo dados semiestruturados relacionados com pacientes, patologias, procedimentos e equipamentos. A exploração destes repositórios para fins de investigação e inteligência empresarial, tem potencial para melhorar os padrões de qualidade e eficiência da prática clínica. No entanto, estes cenários avançados são difíceis de acomodar na realidade atual dos sistemas e redes institucionais. O pobre desempenho de alguns protocolos standard usados em ambiente de produção, conduziu ao uso de soluções proprietárias nestes nichos aplicacionais, limitando a interoperabilidade de sistemas e a integração de fontes de dados. Este doutoramento investigou, desenvolveu e propõe um conjunto de métodos computacionais cujo objetivo é maximizar o desempenho das atuais redes de imagem médica em serviços de pesquisa e recuperação de conteúdos, promovendo a sua utilização em ambientes de elevados requisitos aplicacionais. As propostas foram instanciadas sobre uma plataforma de código aberto e espera-se que ajudem a promover o seu uso generalizado como solução vendor-neutral. As metodologias foram ainda instanciadas e validadas em cenários de uso avançado. Finalmente, é expectável que o trabalho desenvolvido possa facilitar a investigação em ambiente hospitalar de produção, promovendo, desta forma, um aumento da qualidade e eficiência dos serviços.Programa Doutoral em Engenharia Informátic
    • …
    corecore