66 research outputs found

    Txt2vz: a new tool for generating graph clouds

    Get PDF
    We present txt2vz (txt2vz.appspot.com), a new tool for automatically generating a visual summary of unstructured text data found in documents or web sites. The main purpose of the tool is to give the user information about the text so that they can quickly get a good idea about the topics covered. Txt2vz is able to identify important concepts from unstructured text data and to reveal relationships between those concepts. We discuss other approaches to generating diagrams from text and highlight the differences between tag clouds, word clouds, tree clouds and graph clouds

    Looking at a digital research data archive - Visual interfaces to EASY

    Full text link
    In this paper we explore visually the structure of the collection of a digital research data archive in terms of metadata for deposited datasets. We look into the distribution of datasets over different scientific fields; the role of main depositors (persons and institutions) in different fields, and main access choices for the deposited datasets. We argue that visual analytics of metadata of collections can be used in multiple ways: to inform the archive about structure and growth of its collection; to foster collections strategies; and to check metadata consistency. We combine visual analytics and visual enhanced browsing introducing a set of web-based, interactive visual interfaces to the archive's collection. We discuss how text based search combined with visual enhanced browsing enhances data access, navigation, and reuse.Comment: Submitted to the TPDL 201

    Scalable Programming for the Analysis of Aphasia Transcripts

    Get PDF
    Technologies designed for individuals with nonfluent aphasia focus on digitizing speech therapy methods and generating speech. To improve these technologies, the language characteristics of individuals with non- fluent aphasia must be further understood. Language corpuses, such as the AphasiaBank, provide a promising solution for informing technology usability in terms of navigation, interface, and content decisions. As a tool for informing such work, this research investigates the viability of a flexible and scalable multi-threaded software program for the analysis of AphasiaBank transcripts. Results show that the program allows rapid analysis of all transcriptions by optimizing core functionality and minimizing the number of areas for synchronization. This research aims to improve the access to information, products, and services in technology for individuals with non-fluent aphasia

    A Picture is Worth 150 Words: Using Wordle to Assess Library Instruction

    Get PDF
    Tired of the one minute paper and other quick and dirty assessment tools? By using word clouds, students can demonstrate their grasp of library fundamentals and information literacy concepts in less than 10 minutes. Wordle [http://www.wordle.net] is an extremely user-friendly online tool that provides an active learning activity for students and allows librarians to rapidly evaluate what students recall from the instruction session. Use it for quick assessment of student comprehension of library jargon or compare the students\u27 Wordle clouds with information literacy standards or the main points of your instruction. It\u27s free, flexible, and looks great on a t-shirt

    On Semantic Word Cloud Representation

    Full text link
    We study the problem of computing semantic-preserving word clouds in which semantically related words are close to each other. While several heuristic approaches have been described in the literature, we formalize the underlying geometric algorithm problem: Word Rectangle Adjacency Contact (WRAC). In this model each word is associated with rectangle with fixed dimensions, and the goal is to represent semantically related words by ensuring that the two corresponding rectangles touch. We design and analyze efficient polynomial-time algorithms for some variants of the WRAC problem, show that several general variants are NP-hard, and describe a number of approximation algorithms. Finally, we experimentally demonstrate that our theoretically-sound algorithms outperform the early heuristics

    Customer Feedback Analysis using Collocations

    Get PDF
    Today’s ERP and CRM systems provide companies with nearly unlimited possibilities for collecting data concerning theircustomers. More and more of these data are more or less unstructured textual data. A good example of this type of data iscustomer feedback, which can potentially be used to improve customer satisfaction.However, merely getting an overview of what lies in an unstructured mass of text is an extremely challenging task. This isthe topic of the field of computational linguistics. Collocation analysis, one of the tools emerging from this field, is a tooldeveloped for this task in particular. In this paper, we use the collocation analysis to study a text corpora consisting of 64,806pieces of customer feedback collected through a case company’s online customer portal. Collocation analysis is shown to bea very useful tool for exploratory analysis of highly unstructured customer feedback
    • …
    corecore