972 research outputs found

    Sixth Annual Users' Conference

    Get PDF
    Conference papers and presentation outlines which address the use of the Transportable Applications Executive (TAE) and its various applications programs are compiled. Emphasis is given to the design of the user interface and image processing workstation in general. Alternate ports of TAE and TAE subsystems are also covered

    Data processing model for the CDF experiment

    Get PDF
    The data processing model for the CDF experiment is described. Data processing reconstructs events from parallel data streams taken with different combinations of physics event triggers and further splits the events into datasets of specialized physics datasets. The design of the processing control system faces strict requirements on bookkeeping records, which trace the status of data files and event contents during processing and storage. The computing architecture was updated to meet the mass data flow of the Run II data collection, recently upgraded to a maximum rate of 40 MByte/sec. The data processing facility consists of a large cluster of Linux computers with data movement managed by the CDF data handling system to a multi-petaByte Enstore tape library. The latest processing cycle has achieved a stable speed of 35 MByte/sec (3 TByte/day). It can be readily scaled by increasing CPU and data-handling capacity as required.Comment: 12 pages, 10 figures, submitted to IEEE-TN

    Special Libraries, Winter 1987

    Get PDF
    Volume 78, Issue 1https://scholarworks.sjsu.edu/sla_sl_1987/1000/thumbnail.jp

    Integrated Online Media Management Systems For Media Centers: A Model For Selection And Effective Use

    Get PDF
    The researcher proposed to advice in the selection of an Integrated Online Library System (IOLS) for use in the 103 school media centers in the Palm Beach County Schools, Florida. This was accomplished by evaluating the two finalists of those vendors who answered the district\u27s Request for Proposal (RFP). Of the five vendors who responded to the RFP. CLSI and SIRSI were selected as the systems most likely to meet the needs of the media centers of the school district. An overview and definition of IOLS was first discussed. This overview then related itself to the needs of the school district as presented in the RFP. A selection criteria was then designed from previous research on the subject to help find the ideal system. The history and development of an Integrated Online Library System was important in seeing where the systems have originated in contrast to the systems in the Eighties to the also revealed the present time. Literature IOLS principles of operation. The Request for Proposal reflected the needs assessment discussed over several years of committee meetings of representatives from various schools. The committees explored IOLS automated options and compared these options. The RFP outlined the system requirements. Thoughts on staff attitudes while planning for a system were also considered. Each system was evaluated with the criteria outlined in the RFP. The background and capabilities of both systems were explored. This exploration took place where by the benchmark tests, on-site demonstrations systems were in use daily, conferences with the vendors, systems. And reading literature reviews on both systems. Evaluation guidelines and criteria were found in library resources. The functions required for terminal access requirements, process for data conversion, vendor background and reliability, contained in these library resources. And the cost were contained in these library resources. The results of this study culminated in the official recommendation of the SIRSI to be purchased by the district\u27s school board. It was the expectation of the author of this document to see the purchase of the recommended system by the school board and have it implemented in all the schools in the district within a three year period following the submission of the recommendation

    In the Beginning... A Legacy of Computing at Marshall University

    Get PDF
    This book provides a brief history of the early computing technology at Marshall University, Huntington, W.Va., in the forty years: 1959-1999. This was before the move to Intel and Windows based servers. After installation of an IBM Accounting Machine in 1959, which arguably does not fit the modern definition of a computer, the first true computer arrived in 1963 and was installed in a room below the Registrar’s office. For the next twenty years several departments ordered their own midrange standalone systems to fit their individual departmental requirements. These represented different platforms from different vendors, and were not connected to each other. At the same time, the Marshall Computer Center developed an interconnected, multi-processor environment. With the software problems of year 2000, and the I/T move to the new Drinko Library, several systems were scrapped. New systems were installed on the pc server platforms. This book includes images of the various systems, several comments from users, and hardware and software descriptions

    A guide to the National Space Science Data Center

    Get PDF
    This is the second edition of a document that was published to acquaint space and Earth research scientists with an overview of the services offered by the NSSDC. As previously stated, the NSSDC was established by NASA to be the long term archive for data from its space missions. However, the NSSDC has evolved into an organization that provides a multitude of services for scientists throughout the world. Brief articles are presented which discuss these services. At the end of each article is the name, address, and telephone number of the person to contact for additional information. Online Information and Data Systems, Electronic Access, Offline Data Archive, Value Added Services, Mass Storage Activities, and Computer Science Research are all detailed

    Deep Space Network information system architecture study

    Get PDF
    The purpose of this article is to describe an architecture for the Deep Space Network (DSN) information system in the years 2000-2010 and to provide guidelines for its evolution during the 1990s. The study scope is defined to be from the front-end areas at the antennas to the end users (spacecraft teams, principal investigators, archival storage systems, and non-NASA partners). The architectural vision provides guidance for major DSN implementation efforts during the next decade. A strong motivation for the study is an expected dramatic improvement in information-systems technologies, such as the following: computer processing, automation technology (including knowledge-based systems), networking and data transport, software and hardware engineering, and human-interface technology. The proposed Ground Information System has the following major features: unified architecture from the front-end area to the end user; open-systems standards to achieve interoperability; DSN production of level 0 data; delivery of level 0 data from the Deep Space Communications Complex, if desired; dedicated telemetry processors for each receiver; security against unauthorized access and errors; and highly automated monitor and control

    HSLIC Annual Report FY1992-93

    Get PDF
    https://digitalrepository.unm.edu/hslic-annual-reports/1024/thumbnail.jp

    On-line data archives

    Get PDF
    ©2001 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.Digital libraries and other large archives of electronically retrievable and manipulable material are becoming widespread in both commercial and scientific arenas. Advances in networking technologies have led to a greater proliferation of wide-area distributed data warehousing with associated data management challenges. We review tools and technologies for supporting distributed on-line data archives and explain our key concept of active data archives, in which data can be, processed on-demand before delivery. We are developing wide-area data warehousing software infrastructure for geographically distributed archives of large scientific data sets, such as satellite image data, that are stored hierarchically on disk arrays and tape silos and are accessed by a variety of scientific and decision support applications. Interoperability is a major issue for distributed data archives and requires standards for server interfaces and metadata. We review present activities and our contributions in developing such standards for different application areas.K. Hawick, P. Coddington, H. James, C. Patte
    • …
    corecore