13 research outputs found

    Using Latent Semantic Analysis to Identify Themes in IS Healthcare Research

    Get PDF
    Latent Semantic Analysis (LSA) is a new text mining approach that is increasingly being adopted by IS scholars. In this paper, we provide an overview of various research contexts in which IS and other business scholars have applied this approach. We first identity the diverse body of scholarly and field-based contexts in which LSA has been applied. Next, we propose an empirical analysis of published research on healthcare information technology (HIT), to identify different themes in the IS literature from 1990 to the present date. Our empirical analysis will identify key research trends in IS journal papers for three time periods, based on an analysis of health-related papers published in 20 leading IS journals. In addition to providing more awareness of this research approach, we seek to identify important trends and changes in healthcare IT research over time

    New frontiers for qualitative textual data analysis: a multimethod statistical approach

    Get PDF
    In recent years, the increase in textual data production has meant that researchers require faster text analysis techniques and software to reliably produce knowledge for the scientific-nursing community. Automatic text data analysis opens the frontiers to a new research area combining the depth of analysis typical of qualitative research and the stability of measurements required for quantitative studies. Thanks to the statistical-computational approach, it proposes to study more or less extensive written texts produced in natural language to reveal lexical and linguistic worlds and extract useful and meaningful information for researchers. This article aims to provide an overview of this methodology, which has been rarely used in the nursing community to date

    Semantic code search and analysis

    Get PDF
    Title from PDF of title page, viewed on July 28, 2014Thesis advisor: Yugyung LeeVitaIncludes bibliographical references (pages 33-35)Thesis (M. S.)--School of Computing and Engineering. University of Missouri--Kansas City, 2014As open source software repositories have been enormously growing, the high quality source codes have been widely available. A greater access to open source software also leads to an increase of software quality and reduces the overhead of software development. However, most of the available search engines are limited to lexical or code based searches and do not take semantics that underlie the source codes. Thus, object oriented (OO) principles, such as inheritance and composition, cannot be efficiently utilized for code search or analysis. This thesis proposes a novel approach for searching source code using semantics and structure. This approach will allow users to analyze software systems in terms of code similarity. For this purpose, a semantic measurement, called CoSim, was designed based on OO programing models including Package, Class, Method and Interface. We accessed and extracted the source code from open source repositories like Github and converted them into Resource Description Framework (RDF) model. Using the measurement, we queried the source code with SPARQL Query Language and analyzed the systems. We carried out a pilot study for preliminary evaluation of seven different versions of Apache Hadoop systems in terms of their similarities. In addition, we compared the search outputs from our system with those by the Github Code Search. It was shown that our search engine provided more comprehensive and relevant information than the Github does. In addition, the proposed CoSim measurement precisely reflected the significant and evolutionary properties of the systems in the similarity comparison of Hadoop software systemsAbstract -- Illustrations -- Tables - Introduction -- Background and related work -- Semantic code search and analysis model -- Semantic code search and analysis implementation -- Results and evaluation -- Conclusion and future work -- Reference

    funcGNN: A Graph Neural Network Approach to Program Similarity

    Full text link
    Program similarity is a fundamental concept, central to the solution of software engineering tasks such as software plagiarism, clone identification, code refactoring and code search. Accurate similarity estimation between programs requires an in-depth understanding of their structure, semantics and flow. A control flow graph (CFG), is a graphical representation of a program which captures its logical control flow and hence its semantics. A common approach is to estimate program similarity by analysing CFGs using graph similarity measures, e.g. graph edit distance (GED). However, graph edit distance is an NP-hard problem and computationally expensive, making the application of graph similarity techniques to complex software programs impractical. This study intends to examine the effectiveness of graph neural networks to estimate program similarity, by analysing the associated control flow graphs. We introduce funcGNN, which is a graph neural network trained on labeled CFG pairs to predict the GED between unseen program pairs by utilizing an effective embedding vector. To our knowledge, this is the first time graph neural networks have been applied on labeled CFGs for estimating the similarity between high-level language programs. Results: We demonstrate the effectiveness of funcGNN to estimate the GED between programs and our experimental analysis demonstrates how it achieves a lower error rate (0.00194), with faster (23 times faster than the quickest traditional GED approximation method) and better scalability compared with the state of the art methods. funcGNN posses the inductive learning ability to infer program structure and generalise to unseen programs. The graph embedding of a program proposed by our methodology could be applied to several related software engineering problems (such as code plagiarism and clone identification) thus opening multiple research directions.Comment: 11 pages, 8 figures, 3 table

    Latent Problem Solving Analysis as an explanation of expertise effects in a complex, dynamic task

    Get PDF
    Abstract Latent Problem Solving Analysis (LPSA) is a theory of knowledge representation in complex problem solving that argues that problem spaces can be represented as multidimensional spaces and expertise is the construction of those spaces from immense amounts of experience. The model was applied using a dataset from a longitudinal experiment on control of thermodynamic systems. When the system is trained with expert-level amounts of experience (3 years), it can predict the end of a trial using the first three quarters with an accuracy of .9. If the system is prepared to mimic a novice (6 months) the prediction accuracy falls to .2. If the system is trained with 3 years of practice in an environment with no constraints, performance is similar to the novice baseline

    A Review of the Literature on the Empathy Construct Using Cluster Analysis

    Get PDF
    Empathy plays a central role in human behavior and is a key aspect of social functioning. The extensive research on the empathy construct in fields such as psychology, social work, and education has revealed many positive aspects of empathy. Through the use of cluster analysis, this research takes a new approach to reviewing the literature on empathy and objectively identifies groups of empathy research. Next, this study relates the information systems (IS) discipline’s focus on empathy research through the projection of IS empathy paragraphs into those clusters, and identifies areas of empathy research that are currently being largely overlooked by the IS field. The use of cluster analysis and projection for conducting a literature review provides researchers with a more objective approach for reviewing relevant literature

    Using reconfigurable computing technology to accelerate matrix decomposition and applications

    Get PDF
    Matrix decomposition plays an increasingly significant role in many scientific and engineering applications. Among numerous techniques, Singular Value Decomposition (SVD) and Eigenvalue Decomposition (EVD) are widely used as factorization tools to perform Principal Component Analysis for dimensionality reduction and pattern recognition in image processing, text mining and wireless communications, while QR Decomposition (QRD) and sparse LU Decomposition (LUD) are employed to solve the dense or sparse linear system of equations in bioinformatics, power system and computer vision. Matrix decompositions are computationally expensive and their sequential implementations often fail to meet the requirements of many time-sensitive applications. The emergence of reconfigurable computing has provided a flexible and low-cost opportunity to pursue high-performance parallel designs, and the use of FPGAs has shown promise in accelerating this class of computation. In this research, we have proposed and implemented several highly parallel FPGA-based architectures to accelerate matrix decompositions and their applications in data mining and signal processing. Specifically, in this dissertation we describe the following contributions: • We propose an efficient FPGA-based double-precision floating-point architecture for EVD, which can efficiently analyze large-scale matrices. • We implement a floating-point Hestenes-Jacobi architecture for SVD, which is capable of analyzing arbitrary sized matrices. • We introduce a novel deeply pipelined reconfigurable architecture for QRD, which can be dynamically configured to perform either Householder transformation or Givens rotation in a manner that takes advantage of the strengths of each. • We design a configurable architecture for sparse LUD that supports both symmetric and asymmetric sparse matrices with arbitrary sparsity patterns. • By further extending the proposed hardware solution for SVD, we parallelize a popular text mining tool-Latent Semantic Indexing with an FPGA-based architecture. • We present a configurable architecture to accelerate Homotopy l1-minimization, in which the modification of the proposed FPGA architecture for sparse LUD is used at its core to parallelize both Cholesky decomposition and rank-1 update. Our experimental results using an FPGA-based acceleration system indicate the efficiency of our proposed novel architectures, with application and dimension-dependent speedups over an optimized software implementation that range from 1.5ÃÂ to 43.6ÃÂ in terms of computation time
    corecore