111 research outputs found

    Data, Information, and Knowledge in Visualization

    Get PDF
    This article proposes to help advance visualization technology from the interactive visualization of today to the knowledge-assisted visualization of tomorrow by defining data, information and knowledge in the context of visualization

    What May Visualization Processes Optimize?

    Full text link
    In this paper, we present an abstract model of visualization and inference processes and describe an information-theoretic measure for optimizing such processes. In order to obtain such an abstraction, we first examined six classes of workflows in data analysis and visualization, and identified four levels of typical visualization components, namely disseminative, observational, analytical and model-developmental visualization. We noticed a common phenomenon at different levels of visualization, that is, the transformation of data spaces (referred to as alphabets) usually corresponds to the reduction of maximal entropy along a workflow. Based on this observation, we establish an information-theoretic measure of cost-benefit ratio that may be used as a cost function for optimizing a data visualization process. To demonstrate the validity of this measure, we examined a number of successful visualization processes in the literature, and showed that the information-theoretic measure can mathematically explain the advantages of such processes over possible alternatives.Comment: 10 page

    Software-Assisted Knowledge Generation in the Archaeological Domain: A Conceptual Framework

    Get PDF
    Comunicación presentada en la 25th International Conference on Advanced Information Systems Engineering (CAiSE 2013), celebrada en Valencia del 17 al 21 de junio de 2013.Knowledge generation processes are traditionally related to the DIKW (data-information-knowledge-wisdom) hierarchy, a layered model for the classification of human understanding. Software components can be situated in one or several of these layers, or assist in the interfaces between two of them. Most of the knowledge generation processes that occur in the archaeology field involve complex mechanisms of abstraction, relation and interpretation. Is it possible to assist the users in performing these processes? We have detected problems in the archaeological knowledge generation process that could be improved through software assistance. We propose a conceptual framework based on the structure of the data that is being managed by the user, and on the cognitive processes that the user wishes to perform on the data. The proposed framework can, arguably, set the foundation for assisted knowledge generation implemented as software systems.Peer Reviewe

    The role of data visualization in Railway Big Data Risk Analysis

    Get PDF
    Big Data Risk Analysis (BDRA) is one of the possible alleys for the further development of risk models in the railway transport. Big Data techniques allow a great quantity of information to be handled from different types of sources (e.g. unstructured text, signaling and train data). The benefits of this approach may lie in improving the understanding of the risk factors involved in railways, detecting possible new threats or assessing the risk levels for rolling stock, rail infrastructure or railway operations. For the efficient use of BDRA, the conversion of huge amounts of data into a simple and effective display is particularly challenging. Especially because it is presented to various specific target audiences. This work reports a literature review of risk communication and visualization in order to find out its applicability to BDRA, and beyond the visual techniques, what human factors have to be considered in the understanding and risk perception of the infor-mation when safety analysts and decision-makers start basing their decisions on BDRA analyses. It was found that BDRA requires different visualization strategies than those that have normally been carried out in risk analysis up to now

    Mineração, estruturação e disseminação de conhecimento especializado em engenharia de petróleo

    Get PDF
    Orientadores: Celso Kazuyuki Morooka, Kazuo MiuraDissertação (mestrado) - Universidade Estadual de Campinas, Faculdade de Engenharia Mecânica e Instituto de GeociênciasResumo: A aquisição e transmissão de conhecimento são tarefas essenciais que todos os indivíduos e empresas devem enfrentar para subsistir e progredir. Na indústria do petróleo grandes quantidades de textos são estruturados diariamente para facilitar a disseminação de conhecimento, mas o ser humano não tem a habilidade de ler, compreender e lembrar tal quantidade de informação sem ajuda de sistemas computadorizados. Com o propósito de promover a disseminação de conhecimento sobre a engenharia petrolífera a dissertação propõe uma metodologia que permite a aquisição e a disseminação do conhecimento. A metodologia permite extrair os conhecimentos contidos em documentos textuais e mostrá-los graficamente usando mineração de textos e técnicas de visualização. Tal metodologia foi aplicada em duas bases de dados que são Alertas de Segurança da BSEE e teses de doutorado e dissertações de mestrado da UNICAMP as considerando repletas de conhecimento para a indústria de petróleo. A metodologia foi aplicada duas vezes na base de dados da BSEE. A primeira vez para conhecer o conteúdo geral e a segunda para especializar o conhecimento sobre a construção de poços. Os resultados obtidos são "conceitos relevantes" referentes à construção de poços sobre os quais foram construídas três estruturas de conhecimento. Estas estruturas evidenciam as relações existentes e a relevância desses conceitos. Os modelos de conhecimento estruturado obtidos podem ser utilizados para disseminar conhecimento, classificar lições aprendidas, treinar pessoal, visualizar e navegar em conteúdo. O resultado principal desta aplicação é um grafo de conhecimento multicamada que permite a busca por conteúdo e a eficiente recuperação de documentos. A qualidade dos resultados oriundos desta metodologia foram confirmados através de dois testes. O primeiro teste consistiu em buscar dentro da base de dados da UNICAMP, documentos relevantes para estudantes do programa de pós graduação em ciências e engenharia de petróleo (CEP) que estavam realizando trabalhos em diferentes linhas de pesquisa. O segundo teste incidiu em encontrar Alertas de Segurança utilizando palavras chaves idênticas por diferentes motores de busca (motor de busca da BSEE, Google e o método proposto). Os resultados obtidos em ambos os testes mostram a efetividade da metodologia proposta em processar bases de dados locais e especializadasAbstract: Acquisition and transmission of knowledge are essential tasks that all individual and enterprises face to subsist and progress. In the petroleum industry large amounts of texts are daily structured to facilitate the dissemination of knowledge but the human being does not have the ability to read, comprehend and remember such amount of information without the help of computerized systems. With the purpose of promoting the dissemination of knowledge about the petroleum engineering the dissertation proposes a methodology that allows acquisition and dissemination of knowledge. The methodology enables to extract the knowledge contained in textual documents and illustrates it in a graphical format, using text mining and visualization techniques. Such methodology has been applied in two databases, BSEE¿s Safety Alerts and doctoral thesis and master dissertations from CEP-UNICAMP, considering them meaningful sources of knowledge for petroleum industry. On BSEE's database, the methodology has been applied twice. The first time to notice the general content and the second time to specialize the knowledge on well construction. The results obtained are "relevant concepts" about well construction, with which were built three structures of knowledge. Those structures display the relevance and relationship between concepts and can be useful to disseminate knowledge, classify learned lessons, train personnel, visualize and navigate on content. The main result of application is a "Multilayer Knowledge Graph" that allows the research for contents and efficient documents recovery. The quality of results provided by the methodology were confirmed by two tests. The first test consisted to find relevant documents to graduate students of the CEP (Graduate program in petroleum science and engineering) from UNICAMP's database, who were carrying out works in different lines of research. The second test consisted to find Safety Alerts by using identical keywords but different search engines (BSEE's search engine, Google and the proposed method). Results obtained from both tests demonstrated the effectiveness of the proposed methodology in processing local and specialized databasesMestradoExplotaçãoMestra em Ciências e Engenharia de Petróleo190568/2013-5CNP

    Reducing Complex Visualizations for Analysis

    Get PDF
    Data visualization provides a means to present known information in a format that is easily consumable and does not generally require specialized training. It is also well-suited to aid an analyst in discovering previously unknown information [1]. This is possible because visualization techniques can be used to highlight internal relationships and structures within the data, and present them in a graphical manner. Using visualization during the preliminary analysis phase can provide a pathway to enable an analyst to discover patterns or anomalies within the data that might otherwise go undiscovered as humans have an innate ability to visually identify patterns and anomalies. \ \ Even when an analyst has identified a pattern or anomaly within the data, creating an algorithm that allows for automated detection of other occurrences of the same, or similar, patterns is a non-trivial task. While humans are innately skilled at pattern recognition, computers are not, and patterns that might be obvious for a human to identify might be difficult for a computer to detect even when assisted by a skilled analyst [2]. This paper describes a method of taking a complex visualization, and reducing it into several smaller components in order to facilitate computer analysis of the analyst-identified patterns or anomalies in the data. From there, a detection scheme can be generated through an analyst-supervised data analysis process in order to find more occurrences in a larger dataset

    Visualization Criteria: supporting knowledge transfer in Incident

    Get PDF
    Incident Management Systems (IMS) assist in managing resources in order to minimize fatalities and damage. Visual artifacts in an IMS can facilitate knowledge transfer between responders to an incident, however, evidence-based guidance on the design of these visualizations are lacking. The aim of this study is to propose evidence-based knowledge visualization criteria (KVC). Design Science Research (DSR) was the guiding methodology. We abstracted a set of KVC from the academic literature, and then applied said criteria to evaluate a cloud-based prototype IMS. The evaluation included interviews with content experts from the South African Fire Service to establish the relevance of the KVC. The KVC were also used in a heuristic evaluation of the IMS by usability experts. The theoretical contribution of the study is the validated set of KVC based on the triangulation of the findings from the content experts and the usability experts. The study also makes a practical contribution by demonstrating the use of evidencebased visualization criteria in IMS.School of Computin

    A system for exploring big data: an iterative k-means searchlight for outlier detection on open health data

    Full text link
    The interactive exploration of large and evolving datasets is challenging as relationships between underlying variables may not be fully understood. There may be hidden trends and patterns in the data that are worthy of further exploration and analysis. We present a system that methodically explores multiple combinations of variables using a searchlight technique and identifies outliers. An iterative k-means clustering algorithm is applied to features derived through a split-apply-combine paradigm used in the database literature. Outliers are identified as singleton or small clusters. This algorithm is swept across the dataset in a searchlight manner. The dimensions that contain outliers are combined in pairs with other dimensions using a susbset scan technique to gain further insight into the outliers. We illustrate this system by anaylzing open health care data released by New York State. We apply our iterative k-means searchlight followed by subset scanning. Several anomalous trends in the data are identified, including cost overruns at specific hospitals, and increases in diagnoses such as suicides. These constitute novel findings in the literature, and are of potential use to regulatory agencies, policy makers and concerned citizens.Comment: 2018 International Joint Conference on Neural Networks (IJCNN
    corecore