17,836 research outputs found
Impliance: A Next Generation Information Management Appliance
ably successful in building a large market and adapting to the changes of the
last three decades, its impact on the broader market of information management
is surprisingly limited. If we were to design an information management system
from scratch, based upon today's requirements and hardware capabilities, would
it look anything like today's database systems?" In this paper, we introduce
Impliance, a next-generation information management system consisting of
hardware and software components integrated to form an easy-to-administer
appliance that can store, retrieve, and analyze all types of structured,
semi-structured, and unstructured information. We first summarize the trends
that will shape information management for the foreseeable future. Those trends
imply three major requirements for Impliance: (1) to be able to store, manage,
and uniformly query all data, not just structured records; (2) to be able to
scale out as the volume of this data grows; and (3) to be simple and robust in
operation. We then describe four key ideas that are uniquely combined in
Impliance to address these requirements, namely the ideas of: (a) integrating
software and off-the-shelf hardware into a generic information appliance; (b)
automatically discovering, organizing, and managing all data - unstructured as
well as structured - in a uniform way; (c) achieving scale-out by exploiting
simple, massive parallel processing, and (d) virtualizing compute and storage
resources to unify, simplify, and streamline the management of Impliance.
Impliance is an ambitious, long-term effort to define simpler, more robust, and
more scalable information systems for tomorrow's enterprises.Comment: This article is published under a Creative Commons License Agreement
(http://creativecommons.org/licenses/by/2.5/.) You may copy, distribute,
display, and perform the work, make derivative works and make commercial use
of the work, but, you must attribute the work to the author and CIDR 2007.
3rd Biennial Conference on Innovative Data Systems Research (CIDR) January
710, 2007, Asilomar, California, US
Recommended from our members
'Too far ahead of its time': Britain, Burroughs and real-time banking in the 1960s
In 1969, the popular British television programme, Tomorrow's World, featured an item that predicted point of sale terminals in every high street shop ushering in the country's computerised cashless economy. The basis for the show's prediction was a succession of ambitious projects initiated by the British banks, each with the aim of introducing a new real-time computer banking system to its network of branches by 15 February 1971. The banks, threatened by state-sponsored competition, inspired by the success of SABRE, American Airlines' real-time airline reservations system, and pressured by forthcoming decimalisation, all chose 'D-Day' as their immovable deadline. And, in each case, US computer manufacturer, Burroughs, promised a B8500 central super computer linked to a nationwide network of TC500 intelligent terminal satellites. Perhaps unsurprisingly for Tomorrow's World, the programme's predicted coming of the cashless society was wildly optimistic. But so too, it turned out, were the plans of the banks. Real-time banking in Britain never materialised in the 1970s, let alone by February 1971, as one by one the banks abandoned their plans.
In this paper, I revisit the case of Burroughs and Barclays Bank. Blending oral testimonies with archival sources, I explore a consumer perspective of coterminous computing labour as the two companies set about making the idea of real-time banking a reality. I reveal how a community of practice made up of Barclays' computer programmers and Burroughs' engineers was able to transgress established business boundaries in pursuit of a technical ideal, only to eventually become architect of its own fate. The co-construction and 'interpretive flexibility' of this technological failure is considered in light of the existing literature, with particular attention given to the attribution of blame. In this case, where there was attribution, it was judged to have lain with the technology, which was simply regarded as 'too far ahead of its time.ï¿
Computer systems
In addition to the discussions, Ocean Climate Data Workshop hosts gave participants an opportunity to hear about, see, and test for themselves some of the latest computer tools now available for those studying climate change and the oceans. Six speakers described computer systems and their functions. The introductory talks were followed by demonstrations to small groups of participants and some opportunities for participants to get hands-on experience. After this familiarization period, attendees were invited to return during the course of the Workshop and have one-on-one discussions and further hands-on experience with these systems. Brief summaries or abstracts of introductory presentations are addressed
Corporate influence and the academic computer science discipline. [4: CMU]
Prosopographical work on the four major centers for computer
research in the United States has now been conducted, resulting in big
questions about the independence of, so called, computer science
NSSDC Conference on Mass Storage Systems and Technologies for Space and Earth Science Applications, volume 1
Papers and viewgraphs from the conference are presented. This conference served as a broad forum for the discussion of a number of important issues in the field of mass storage systems. Topics include magnetic disk and tape technologies, optical disks and tape, software storage and file management systems, and experiences with the use of a large, distributed storage system. The technical presentations describe, among other things, integrated mass storage systems that are expected to be available commercially. Also included is a series of presentations from Federal Government organizations and research institutions covering their mass storage requirements for the 1990's
A comparison of airborne and ground-based radar observations with rain gages during the CaPE experiment
The vicinity of KSC, where the primary ground truth site of the Tropical Rainfall Measuring Mission (TRMM) program is located, was the focal point of the Convection and Precipitation/Electrification (CaPE) experiment in Jul. and Aug. 1991. In addition to several specialized radars, local coverage was provided by the C-band (5 cm) radar at Patrick AFB. Point measurements of rain rate were provided by tipping bucket rain gage networks. Besides these ground-based activities, airborne radar measurements with X- and Ka-band nadir-looking radars on board an aircraft were also recorded. A unique combination data set of airborne radar observations with ground-based observations was obtained in the summer convective rain regime of central Florida. We present a comparison of these data intending a preliminary validation. A convective rain event was observed simultaneously by all three instrument types on the evening of 27 Jul. 1991. The high resolution aircraft radar was flown over convective cells with tops exceeding 10 km and observed reflectivities of 40 to 50 dBZ at 4 to 5 km altitude, while the low resolution surface radar observed 35 to 55 dBZ echoes and a rain gage indicated maximum surface rain rates exceeding 100 mm/hr. The height profile of reflectivity measured with the airborne radar show an attenuation of 6.5 dB/km (two way) for X-band, corresponding to a rainfall rate of 95 mm/hr
- …