79,420 research outputs found

    Computer Assisted Learning: Its Educational Potential (UNCAL)

    Get PDF

    A transaction-oriented architecture for structuring unstructured information in enterprise applications

    Get PDF
    As 80-85% of all corporate information remains unstructured, outside of the processing scope of enterprise systems, many enterprises rely on Information Systems that cause them to risk transactions that are based on lack of information (errors of omission) or misleading information (errors of commission). To address this concern, the fundamental business concept of monetary transactions is extended to include qualitative business concepts. A Transaction Concept (TC) is accordingly identified that provides a structure for these unstructured but vital aspects of business transactions. Based on REA (Resources, Events, Agents) and modelled using Conceptual Graphs (CGs) and Formal Concept Analysis (FCA), the TC provides businesses with a more balanced view of the transactions they engage in and a means of discovering new transactions that they might have otherwise missed. A simple example is provided that illustrates this integration and reveals a key missing element. This example is supported by reference to a wide range of case studies and application areas that demonstrate the added value of the TC. The TC is then advanced into a Transaction-Oriented Architecture (TOA). The TOA provides the framework by which an enterprise’s business processes are orchestrated according to the TC. TOA thus brings Service-Oriented Architecture (SOA) and the productivity of enterprise applications to the height of the real, transactional world that enterprises actually operate in.</jats:p

    Teaching Data Science

    Get PDF
    We describe an introductory data science course, entitled Introduction to Data Science, offered at the University of Illinois at Urbana-Champaign. The course introduced general programming concepts by using the Python programming language with an emphasis on data preparation, processing, and presentation. The course had no prerequisites, and students were not expected to have any programming experience. This introductory course was designed to cover a wide range of topics, from the nature of data, to storage, to visualization, to probability and statistical analysis, to cloud and high performance computing, without becoming overly focused on any one subject. We conclude this article with a discussion of lessons learned and our plans to develop new data science courses.Comment: 10 pages, 4 figures, International Conference on Computational Science (ICCS 2016

    Experimental Study of Remote Job Submission and Execution on LRM through Grid Computing Mechanisms

    Full text link
    Remote job submission and execution is fundamental requirement of distributed computing done using Cluster computing. However, Cluster computing limits usage within a single organization. Grid computing environment can allow use of resources for remote job execution that are available in other organizations. This paper discusses concepts of batch-job execution using LRM and using Grid. The paper discusses two ways of preparing test Grid computing environment that we use for experimental testing of concepts. This paper presents experimental testing of remote job submission and execution mechanisms through LRM specific way and Grid computing ways. Moreover, the paper also discusses various problems faced while working with Grid computing environment and discusses their trouble-shootings. The understanding and experimental testing presented in this paper would become very useful to researchers who are new to the field of job management in Grid.Comment: Fourth International Conference on Advanced Computing & Communication Technologies (ACCT), 201

    Knowledge formalization in experience feedback processes : an ontology-based approach

    Get PDF
    Because of the current trend of integration and interoperability of industrial systems, their size and complexity continue to grow making it more difficult to analyze, to understand and to solve the problems that happen in their organizations. Continuous improvement methodologies are powerful tools in order to understand and to solve problems, to control the effects of changes and finally to capitalize knowledge about changes and improvements. These tools involve suitably represent knowledge relating to the concerned system. Consequently, knowledge management (KM) is an increasingly important source of competitive advantage for organizations. Particularly, the capitalization and sharing of knowledge resulting from experience feedback are elements which play an essential role in the continuous improvement of industrial activities. In this paper, the contribution deals with semantic interoperability and relates to the structuring and the formalization of an experience feedback (EF) process aiming at transforming information or understanding gained by experience into explicit knowledge. The reuse of such knowledge has proved to have significant impact on achieving themissions of companies. However, the means of describing the knowledge objects of an experience generally remain informal. Based on an experience feedback process model and conceptual graphs, this paper takes domain ontology as a framework for the clarification of explicit knowledge and know-how, the aim of which is to get lessons learned descriptions that are significant, correct and applicable
    • …
    corecore