424 research outputs found
The ingenuity of common workmen: and the invention of the computer
Since World War II, state support for scientific research has been assumed crucial to technological and economic progress. Governments accordingly spent tremendous sums to that end. Nothing epitomizes the alleged fruits of that involvement better than the electronic digital computer. The first such computer has been widely reputed to be the ENIAC, financed by the U.S. Army for the war but finished afterwards. Vastly improved computers followed, initially paid for in good share by the Federal Government of the United States, but with the private sector then dominating, both in development and use, and computers are of major significance.;Despite the supposed success of public-supported science, evidence is that computers would have evolved much the same without it but at less expense. Indeed, the foundations of modern computer theory and technology were articulated before World War II, both as a tool of applied mathematics and for information processing, and the computer was itself on the cusp of reality. Contrary to popular understanding, the ENIAC actually represented a movement backwards and a dead end.;Rather, modern computation derived more directly, for example, from the prewar work of John Vincent Atanasoff and Clifford Berry, a physics professor and graduate student, respectively, at Iowa State College (now University) in Ames, Iowa. They built the Atanasoff Berry Computer (ABC), which, although special purpose and inexpensive, heralded the efficient and elegant design of modern computers. Moreover, while no one foresaw commercialization of computers based on the ungainly and costly ENIAC, the commercial possibilities of the ABC were immediately evident, although unrealized due to war. Evidence indicates, furthermore, that the private sector was willing and able to develop computers beyond the ABC and could have done so more effectively than government, to the most sophisticated machines.;A full and inclusive history of computers suggests that Adam Smith, the eighteenth century Scottish philosopher, had it right. He believed that minimal and aloof government best served society, and that the inherent genius of citizens was itself enough to ensure the general prosperity
Informatics in the Future: Proceedings of the 11th European Computer Science Summit (ECSS 2015), Vienna, October 2015
Big data; Computing ethics; Women in computing; Research ethic
Recommended from our members
The benefit of distance learning
This project will address the issues concerning the benefits of distance learning, a growing area in our education system today. Distance learning takes place when the teacher and students are separated, and technology is used to bridge the instructional gap
Recommended from our members
Inventing Intelligence: On the History of Complex Information Processing and Artificial Intelligence in the United States in the Mid-Twentieth Century
In the mid-1950s, researchers in the United States melded formal theories of problem solving and intelligence with another powerful new tool for control: the electronic digital computer. Several branches of western mathematical science emerged from this nexus, including computer science (1960s–), data science (1990s–) and artificial intelligence (AI). This thesis offers an account of the origins and politics of AI in the mid-twentieth century United States, which focuses on its imbrications in systems of societal control. In an effort to denaturalize the power relations upon which the field came into being, I situate AI’s canonical origin story in relation to the structural and intellectual priorities of the U.S. military and American industry during the Cold War, circa 1952 to 1961.
This thesis offers a detailed and comparative account of the early careers, research interests, and key outputs of four researchers often credited with laying the foundations for AI and machine learning—Herbert A. Simon, Frank Rosenblatt, John McCarthy and Marvin Minsky. It chronicles the distinct ways in which each sought to formalise and simulate human mental behaviour using digital electronic computers. Rather than assess their contributions as discontinuous with what came before, as in mythologies of AI's genesis, I establish continuities with, and borrowings from, management science and operations research (Simon), Hayekian economics and instrumentalist statistics (Rosenblatt), automatic coding techniques and pedagogy (McCarthy), and cybernetics (Minsky), along with the broadscale mobilization of Cold War-era civilian-led military science generally.
I assess how Minsky’s 1961 paper 'Steps Toward Artificial Intelligence' simultaneously consolidated and obscured these entanglements as it set in motion an initial research agenda for AI in the following two decades. I argue that mind-computer metaphors, and research in complex information processing generally, played an important role in normalizing the small- and large-scale structuring of social behaviour using mathematics in the United States from the second half of the twentieth century onward
“INDUSTRIAL LEGISLATURES”: CONSENSUS STANDARDIZATION IN THE SECOND AND THIRD INDUSTRIAL REVOLUTIONS
Consensus standardization is a social process in which technical experts from
public, private, and non-profit sectors negotiate the direction and shape of technological
change. Scholars in a variety of disciplines have recognized the importance of consensus
standards as alternatives to standards that arise through market mechanisms or standards
mandated by regulators. Rather than treating the consensus method as some sort of
timeless organizational form or ever-present alternative to markets or laws, I argue that
consensus standardization is itself a product of history.
In the first two chapters, I explain the origins and growth of consensus standards
bodies between 1880 and 1930 as a reaction to and critique of the existing political
economy of engineering. By considering the standardization process—instead of the
internal dynamics of a particular firm or technology—as the primary category of analysis,
I am able to emphasize the cooperative relations that sustained the American style of
competitive managerial capitalism during the Second Industrial Revolution. In the
remaining four chapters, I examine the processes of network architecture and
standardization in the creation of four communications networks during the twentieth
century: AT&T’s monopoly telephone network, the Internet, digital cellular telephone
networks, and the World Wide Web.
Each of these four networks embodied critiques—always implicit and frequently
explicit—of preceding and competing networks. These critiques, visible both in the
technological design of networks as well as in the institutional design of standard-setting
bodies, reflected the political convictions of successive generations of engineers and network architects. The networks described in this dissertation were thus turning points
in the century-long development of an organizational form. Seen as part of a common
history, they tell the story of how consensus-based institutions became the dominant
mode for setting standards in the Third Industrial Revolution, and created the
foundational standards of the information infrastructures upon which a newly globalized
economy and society—the Network Society—could grow
Towards a Digital Epistemology
This Open Access book explores the concept of digital epistemology. In this context, the digital will not be understood as merely something that is linked to specific tools and objects, but rather as different modes of thought. For example, the digital within the humanities is not just databases and big data, topic modelling and speculative visualizations; nor are the objects limited to computer games, other electronic works, or to literature and art that explicitly relate to computerization or other digital aspects. In what way do digital tools and expressions in the 1960s differ to the ubiquitous systems of our time? What kind of artistic effects does this generate? Is the present theoretical fascination for materiality an effect or a reaction to a digitization? Above all: how can early modern forms such as the cabinets of curiosity, emblem books and the archival principle of pertinence contribute to the analyses of contemporary digital forms
- …