972 research outputs found
Sixth Annual Users' Conference
Conference papers and presentation outlines which address the use of the Transportable Applications Executive (TAE) and its various applications programs are compiled. Emphasis is given to the design of the user interface and image processing workstation in general. Alternate ports of TAE and TAE subsystems are also covered
Data processing model for the CDF experiment
The data processing model for the CDF experiment is described. Data
processing reconstructs events from parallel data streams taken with different
combinations of physics event triggers and further splits the events into
datasets of specialized physics datasets. The design of the processing control
system faces strict requirements on bookkeeping records, which trace the status
of data files and event contents during processing and storage. The computing
architecture was updated to meet the mass data flow of the Run II data
collection, recently upgraded to a maximum rate of 40 MByte/sec. The data
processing facility consists of a large cluster of Linux computers with data
movement managed by the CDF data handling system to a multi-petaByte Enstore
tape library. The latest processing cycle has achieved a stable speed of 35
MByte/sec (3 TByte/day). It can be readily scaled by increasing CPU and
data-handling capacity as required.Comment: 12 pages, 10 figures, submitted to IEEE-TN
Special Libraries, Winter 1987
Volume 78, Issue 1https://scholarworks.sjsu.edu/sla_sl_1987/1000/thumbnail.jp
Integrated Online Media Management Systems For Media Centers: A Model For Selection And Effective Use
The researcher proposed to advice in the selection of an Integrated Online Library System (IOLS) for use in the 103 school media centers in the Palm Beach County Schools, Florida. This was accomplished by evaluating the two finalists of those vendors who answered the district\u27s Request for Proposal (RFP). Of the five vendors who responded to the RFP. CLSI and SIRSI were selected as the systems most likely to meet the needs of the media centers of the school district.
An overview and definition of IOLS was first discussed. This overview then related itself to the needs of the school district as presented in the RFP. A selection criteria was then designed from previous research on the subject to help find the ideal system.
The history and development of an Integrated Online Library System was important in seeing where the systems have originated in contrast to the systems in the Eighties to the also revealed the present time. Literature IOLS principles of operation.
The Request for Proposal reflected the needs assessment discussed over several years of committee meetings of representatives from various schools. The committees explored IOLS automated options and compared these options. The RFP outlined the system requirements. Thoughts on staff attitudes while planning for a system were also considered.
Each system was evaluated with the criteria outlined in the RFP. The background and capabilities of both systems were explored. This exploration took place where by the benchmark tests, on-site demonstrations systems were in use daily, conferences with the vendors, systems. And reading literature reviews on both systems.
Evaluation guidelines and criteria were found in library resources. The functions required for terminal access requirements, process for data conversion, vendor background and reliability, contained in these library resources. And the cost were contained in these library resources.
The results of this study culminated in the official recommendation of the SIRSI to be purchased by the district\u27s school board. It was the expectation of the author of this document to see the purchase of the recommended system by the school board and have it implemented in all the schools in the district within a three year period following the submission of the recommendation
In the Beginning... A Legacy of Computing at Marshall University
This book provides a brief history of the early computing technology at Marshall University, Huntington, W.Va., in the forty years: 1959-1999. This was before the move to Intel and Windows based servers. After installation of an IBM Accounting Machine in 1959, which arguably does not fit the modern definition of a computer, the first true computer arrived in 1963 and was installed in a room below the Registrar’s office. For the next twenty years several departments ordered their own midrange standalone systems to fit their individual departmental requirements. These represented different platforms from different vendors, and were not connected to each other. At the same time, the Marshall Computer Center developed an interconnected, multi-processor environment. With the software problems of year 2000, and the I/T move to the new Drinko Library, several systems were scrapped. New systems were installed on the pc server platforms. This book includes images of the various systems, several comments from users, and hardware and software descriptions
A guide to the National Space Science Data Center
This is the second edition of a document that was published to acquaint space and Earth research scientists with an overview of the services offered by the NSSDC. As previously stated, the NSSDC was established by NASA to be the long term archive for data from its space missions. However, the NSSDC has evolved into an organization that provides a multitude of services for scientists throughout the world. Brief articles are presented which discuss these services. At the end of each article is the name, address, and telephone number of the person to contact for additional information. Online Information and Data Systems, Electronic Access, Offline Data Archive, Value Added Services, Mass Storage Activities, and Computer Science Research are all detailed
Recommended from our members
National Geoscience Data Repository System, Phase II. Final report, January 30, 1995--January 28, 1997
The American Geological Institute (AGI) has completed Phase II of a project to establish a National Geoscience Data Repository System (NGDRS). The project`s primary objectives are to preserve geoscience data in jeopardy of being destroyed and to make that data available to those who have a need to use it in future investigations. These data are available for donation to the public as a result of the downsizing that has occurred in the major petroleum and mining companies in the United States for the past decade. In recent years, these companies have consolidated domestic operations, sold many of their domestic properties and relinquished many of their leases. The scientific data associated with those properties are no longer considered to be useful assets and are consequently in danger of being lost forever. The national repository project will make many of these data available to the geoscience community for the first time. To address this opportunity, AGI sought support from the Department of Energy (DOE) in 1994 to initiate the NGDRS Phase I feasibility study to determine the types and quantity of data that companies would be willing to donate. The petroleum and mining companies surveyed indicated that they were willing to donate approximately five million well logs, one hundred million miles of seismic reflection data, millions of linear feet of core and cuttings, and a variety of other types of scientific data. Based on the positive results of the Phase I study, AGI undertook Phase II of the program in 1995. Funded jointly by DOE and industry, Phase II encompasses the establishment of standards for indexing and cataloging of geoscience data and determination of the costs of transferring data from the private sector to public-sector data repositories. Pilot projects evaluated the feasibility of the project for transfer of different data types and creation of a Web-based metadata supercatalog and browser
Deep Space Network information system architecture study
The purpose of this article is to describe an architecture for the Deep Space Network (DSN) information system in the years 2000-2010 and to provide guidelines for its evolution during the 1990s. The study scope is defined to be from the front-end areas at the antennas to the end users (spacecraft teams, principal investigators, archival storage systems, and non-NASA partners). The architectural vision provides guidance for major DSN implementation efforts during the next decade. A strong motivation for the study is an expected dramatic improvement in information-systems technologies, such as the following: computer processing, automation technology (including knowledge-based systems), networking and data transport, software and hardware engineering, and human-interface technology. The proposed Ground Information System has the following major features: unified architecture from the front-end area to the end user; open-systems standards to achieve interoperability; DSN production of level 0 data; delivery of level 0 data from the Deep Space Communications Complex, if desired; dedicated telemetry processors for each receiver; security against unauthorized access and errors; and highly automated monitor and control
HSLIC Annual Report FY1992-93
https://digitalrepository.unm.edu/hslic-annual-reports/1024/thumbnail.jp
On-line data archives
©2001 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.Digital libraries and other large archives of electronically retrievable and manipulable material are becoming widespread in both commercial and scientific arenas. Advances in networking technologies have led to a greater proliferation of wide-area distributed data warehousing with associated data management challenges. We review tools and technologies for supporting distributed on-line data archives and explain our key concept of active data archives, in which data can be, processed on-demand before delivery. We are developing wide-area data warehousing software infrastructure for geographically distributed archives of large scientific data sets, such as satellite image data, that are stored hierarchically on disk arrays and tape silos and are accessed by a variety of scientific and decision support applications. Interoperability is a major issue for distributed data archives and requires standards for server interfaces and metadata. We review present activities and our contributions in developing such standards for different application areas.K. Hawick, P. Coddington, H. James, C. Patte
- …