102 research outputs found

    Visual network storytelling

    Get PDF
    We love networks! Networks are powerful conceptual tools, encapsulating in a single item multiple affordances for computation (networks as graphs), visualization (networks as maps) and manipulation of data (networks as interfaces). In the field of mathematics, graph theory has been around since Euler’s walk on Königsberg’s bridges (Euler 1736). But it is not until the end of the last century that networks acquired a multidisciplinary popularity. Graph computation is certainly powerful, but it is also very demanding and for many years its advantages remained the privilege of scholars with solid mathematical fundamentals. In the last few decades, however, networks acquired a new set of affordances and reached a larger audience, thanks to the growing availability of tools to design them. Drawn on paper or screen, networks became easier to handle and obtained properties that calculation could not express. Far from being merely aesthetic, the graphical representation of networks has an intrinsic hermeneutic value. Networks can become maps and be read as such. Combining the computation power of graphs with the visual expressivity of maps and the interactivity of computer interface, networks can be used in Exploratory Data Analysis (Tukey, 1977). Navigating through data becomes so fluid that zooming in on a single data-point and out to a landscape of a million traces is just a click away. Increasingly specialized software has been designed to support the exploration of network data. Tools like Pajek (vlado.fmf.uni-lj.si/pub/networks/pajek), NetDraw (sites.google.com/site/ netdrawsoftware), Ucinet (www.analytictech.com/ucinet), Guess (graphexploration.cond.org) and more recently Gephi (gephi.org) have progressively smoothed out the difficulties of graph mathematics, turning a complex mathematical formalism into a more user-friendly point-and-click interface (1) . If visual exploration of networks can output to confirmatory statistics, what about sharing one network exploration with others? We developed Manylines (https://github.com/medialab/manylines), a tool allowing you to share the visual analysis of a network with a wide audience by publishing it on the web. With Manylines, you can not only easily publish a network on the web but also share its exploration by describing the network’s visual key findings. Through a set of examples, we will illustrate how the narrative opportunities of Manylines can contribute to the enunciation of a visual grammar of networks. (1) A simple look at the URLs of the subsequent tools reveals the efforts deployed to make network-manipulation tools user-friendly and thereby available to a larger public

    Enabling entity retrieval by exploiting Wikipedia as a semantic knowledge source

    Get PDF
    This dissertation research, PanAnthropon FilmWorld, aims to demonstrate direct retrieval of entities and related facts by exploiting Wikipedia as a semantic knowledge source, with the film domain as its proof-of-concept domain of application. To this end, a semantic knowledge base concerning the film domain has been constructed with the data extracted/derived from 10,640 Wikipedia pages on films and additional pages on film awards. The knowledge base currently contains 209,266 entities and 2,345,931 entity-centric facts. Both the knowledge base and the corresponding semantic search interface are based on the coherent classification of entities. Entity-centric facts are also consistently represented as tuples. The semantic search interface (http://dlib.ischool.drexel.edu:8080/sofia/PA/) supports multiple types of semantic search functions, which go beyond the traditional keyword-based search function, including the main General Entity Retrieval Query (GERQ) function, which is concerned with retrieving all entities that match the specified entity type, subtype, and semantic conditions and thus corresponds to the main research problem. Two types of evaluation have been performed in order to evaluate (1) the quality of information extraction and (2) the effectiveness of information retrieval using the semantic interface. The first type of evaluation has been performed by inspecting 11,495 film-centric facts concerning 100 films. The results have confirmed high data quality with 99.96% average precision and 99.84% average recall. The second type of evaluation has been performed by conducting an experiment with human subjects. The experiment involved having the subjects perform a retrieval task by using both the PanAnthropon interface and the Internet Movie Database (IMDb) interface and comparing their task performance between the two interfaces. The results have confirmed higher effectiveness of the PanAnthropon interface vs. the IMDb interface (83.11% vs. 40.78% average precision; 83.55% vs. 40.26% average recall). Moreover, the subjects’ responses to the post-task questionnaire indicate that the subjects found the PanAnthropon interface to be highly usable and easily understandable as well as highly effective. The main contribution from this research therefore consists in achieving the set research goal, namely, demonstrating the utility and feasibility of semantics-based direct entity retrieval.Ph.D., Information Studies -- Drexel University, 201

    Semantic Interaction in Web-based Retrieval Systems : Adopting Semantic Web Technologies and Social Networking Paradigms for Interacting with Semi-structured Web Data

    Get PDF
    Existing web retrieval models for exploration and interaction with web data do not take into account semantic information, nor do they allow for new forms of interaction by employing meaningful interaction and navigation metaphors in 2D/3D. This thesis researches means for introducing a semantic dimension into the search and exploration process of web content to enable a significantly positive user experience. Therefore, an inherently dynamic view beyond single concepts and models from semantic information processing, information extraction and human-machine interaction is adopted. Essential tasks for semantic interaction such as semantic annotation, semantic mediation and semantic human-computer interaction were identified and elaborated for two general application scenarios in web retrieval: Web-based Question Answering in a knowledge-based dialogue system and semantic exploration of information spaces in 2D/3D

    A Quadruple-Based Text Analysis System for History and Philosophy of Science

    Get PDF
    abstract: Computational tools in the digital humanities often either work on the macro-scale, enabling researchers to analyze huge amounts of data, or on the micro-scale, supporting scholars in the interpretation and analysis of individual documents. The proposed research system that was developed in the context of this dissertation ("Quadriga System") works to bridge these two extremes by offering tools to support close reading and interpretation of texts, while at the same time providing a means for collaboration and data collection that could lead to analyses based on big datasets. In the field of history of science, researchers usually use unstructured data such as texts or images. To computationally analyze such data, it first has to be transformed into a machine-understandable format. The Quadriga System is based on the idea to represent texts as graphs of contextualized triples (or quadruples). Those graphs (or networks) can then be mathematically analyzed and visualized. This dissertation describes two projects that use the Quadriga System for the analysis and exploration of texts and the creation of social networks. Furthermore, a model for digital humanities education is proposed that brings together students from the humanities and computer science in order to develop user-oriented, innovative tools, methods, and infrastructures.Dissertation/ThesisDoctoral Dissertation Biology 201

    Theories of Informetrics and Scholarly Communication

    Get PDF
    Scientometrics have become an essential element in the practice and evaluation of science and research, including both the evaluation of individuals and national assessment exercises. Yet, researchers and practitioners in this field have lacked clear theories to guide their work. As early as 1981, then doctoral student Blaise Cronin published "The need for a theory of citing" —a call to arms for the fledgling scientometric community to produce foundational theories upon which the work of the field could be based. More than three decades later, the time has come to reach out the field again and ask how they have responded to this call. This book compiles the foundational theories that guide informetrics and scholarly communication research. It is a much needed compilation by leading scholars in the field that gathers together the theories that guide our understanding of authorship, citing, and impact

    Theories of Informetrics and Scholarly Communication

    Get PDF
    Scientometrics have become an essential element in the practice and evaluation of science and research, including both the evaluation of individuals and national assessment exercises. Yet, researchers and practitioners in this field have lacked clear theories to guide their work. As early as 1981, then doctoral student Blaise Cronin published The need for a theory of citing - a call to arms for the fledgling scientometric community to produce foundational theories upon which the work of the field could be based. More than three decades later, the time has come to reach out the field again and ask how they have responded to this call. This book compiles the foundational theories that guide informetrics and scholarly communication research. It is a much needed compilation by leading scholars in the field that gathers together the theories that guide our understanding of authorship, citing, and impact

    The Trouble with Knowing: Wikipedian consensus and the political design of encyclopedic media

    Get PDF
    Encyclopedias are interfaces between knowing and the unknown. They are devices that negotiate the middle ground between incompatible knowledge systems while also performing as dream machines that explore the political outlines of an enlightened society. Building upon the insights from critical feminist theory, media archaeology, and science and technology studies, the dissertation investigates how utopian and impossible desires of encyclopedic media have left a wake of unresolvable epistemological crises. In a 2011 survey of editors of the online encyclopedia Wikipedia, it was reported that 87 per cent of Wikipedians identified as men. This statistic flew in the face of Wikipedias utopian promise that it was an encyclopedia that anyone can edit. Despite the early optimism and efforts to reduce this disparity, Wikipedias parent organization acknowledged its inability to significantly make Wikipedia more equitable. This matter of concern raised two questions: What kinds of knowing subjects is Wikipedia designed to cultivate and what does this conflict over who is included and excluded within Wikipedia tell us about the utopian dreams that are woven into encyclopedic media? This dissertation argues that answering these troubling questions requires an examination of the details of the present, but also the impossible desires that Wikipedia inherited from its predecessors. The analysis of these issues begins with a genealogy of encyclopedias, encyclopedists, encyclopedic aesthetics, and encyclopedisms. It is followed by an archeology of the twentieth century deployment of consensus as an encyclopedic and political program. The third part examines how Wikipedia translated the imaginary ideal of consensus into a cultural technique. Finally, the dissertation mobilizes these analyses to contextualize how consensus was used to limit the dissenting activities of Wikipedia's Gender Gap Task Force. The dissertation demonstrates that the desire and design of encircling knowledge through consensus cultivated Wikipedias gender gap. In this context, if encyclopedic knowledge is to remain politically and culturally significant in the twenty-first century, it is necessary to tell a new story about encyclopedic media. It must be one where an attention to utopian imaginaries, practices, and techniques not only addresses how knowledge is communicated but also enables a sensitivity to the question of who can know

    Repurposing digital traces to organize social attention

    Get PDF
    In the late 1990s, Google pioneered the idea of scraping and repurposing digital traces as a new form of data with which to understand people’s preferences and behaviour. This way of generating empirical sensitivity towards the world can be termed digital methods and the last five years have seen such methods gain influence beyond the field of Internet search. Organizations of different kinds are increasingly mentioning the need to harness the intelligence of ‘big’ digital datasets, and the social sciences have similarly been marked by suggestions to move away from established methods such as surveys and focus groups, and learn from the way Google and other companies have succeeded in turning big datasets into knowledge of social dynamics. By enabling new combinations of data and software and by providing new ways of searching, aggregating, and cross-referencing empirical datasets, it seems probable that the spread of digital methods will re-configure the way organizations, social scientists, and citizens ‘see’ the world in which they live
    corecore