36,058 research outputs found

    Simplifying Deep-Learning-Based Model for Code Search

    Full text link
    To accelerate software development, developers frequently search and reuse existing code snippets from a large-scale codebase, e.g., GitHub. Over the years, researchers proposed many information retrieval (IR) based models for code search, which match keywords in query with code text. But they fail to connect the semantic gap between query and code. To conquer this challenge, Gu et al. proposed a deep-learning-based model named DeepCS. It jointly embeds method code and natural language description into a shared vector space, where methods related to a natural language query are retrieved according to their vector similarities. However, DeepCS' working process is complicated and time-consuming. To overcome this issue, we proposed a simplified model CodeMatcher that leverages the IR technique but maintains many features in DeepCS. Generally, CodeMatcher combines query keywords with the original order, performs a fuzzy search on name and body strings of methods, and returned the best-matched methods with the longer sequence of used keywords. We verified its effectiveness on a large-scale codebase with about 41k repositories. Experimental results showed the simplified model CodeMatcher outperforms DeepCS by 97% in terms of MRR (a widely used accuracy measure for code search), and it is over 66 times faster than DeepCS. Besides, comparing with the state-of-the-art IR-based model CodeHow, CodeMatcher also improves the MRR by 73%. We also observed that: fusing the advantages of IR-based and deep-learning-based models is promising because they compensate with each other by nature; improving the quality of method naming helps code search, since method name plays an important role in connecting query and code

    Repository-Based Software Engineering Program: Working Program Management Plan

    Get PDF
    Repository-Based Software Engineering Program (RBSE) is a National Aeronautics and Space Administration (NASA) sponsored program dedicated to introducing and supporting common, effective approaches to software engineering practices. The process of conceiving, designing, building, and maintaining software systems by using existing software assets that are stored in a specialized operational reuse library or repository, accessible to system designers, is the foundation of the program. In addition to operating a software repository, RBSE promotes (1) software engineering technology transfer, (2) academic and instructional support of reuse programs, (3) the use of common software engineering standards and practices, (4) software reuse technology research, and (5) interoperability between reuse libraries. This Program Management Plan (PMP) is intended to communicate program goals and objectives, describe major work areas, and define a management report and control process. This process will assist the Program Manager, University of Houston at Clear Lake (UHCL) in tracking work progress and describing major program activities to NASA management. The goal of this PMP is to make managing the RBSE program a relatively easy process that improves the work of all team members. The PMP describes work areas addressed and work efforts being accomplished by the program; however, it is not intended as a complete description of the program. Its focus is on providing management tools and management processes for monitoring, evaluating, and administering the program; and it includes schedules for charting milestones and deliveries of program products. The PMP was developed by soliciting and obtaining guidance from appropriate program participants, analyzing program management guidance, and reviewing related program management documents

    Enhanced Integrated Scoring for Cleaning Dirty Texts

    Full text link
    An increasing number of approaches for ontology engineering from text are gearing towards the use of online sources such as company intranet and the World Wide Web. Despite such rise, not much work can be found in aspects of preprocessing and cleaning dirty texts from online sources. This paper presents an enhancement of an Integrated Scoring for Spelling error correction, Abbreviation expansion and Case restoration (ISSAC). ISSAC is implemented as part of a text preprocessing phase in an ontology engineering system. New evaluations performed on the enhanced ISSAC using 700 chat records reveal an improved accuracy of 98% as compared to 96.5% and 71% based on the use of only basic ISSAC and of Aspell, respectively.Comment: More information is available at http://explorer.csse.uwa.edu.au/reference

    Lessons learned: structuring knowledge codification and abstraction to provide meaningful information for learning

    Get PDF
    Purpose – To increase the spread and reuse of lessons learned (LLs), the purpose of this paper is to develop a standardised information structure to facilitate concise capture of the critical elements needed to engage secondary learners and help them apply lessons to their contexts. Design/methodology/approach – Three workshops with industry practitioners, an analysis of over 60 actual lessons from private and public sector organisations and seven practitioner interviews provided evidence of actual practice. Design science was used to develop a repeatable/consistent information model of LL content/structure. Workshop analysis and theory provided the coding template. Situation theory and normative analysis were used to define the knowledge and rule logic to standardise fields. Findings – Comparing evidence from practice against theoretical prescriptions in the literature highlighted important enhancements to the standard LL model. These were a consistent/concise rule and context structure, appropriate emotional language, reuse and control criteria to ensure lessons were transferrable and reusable in new situations. Research limitations/implications – Findings are based on a limited sample. Long-term benefits of standardisation and use need further research. A larger sample/longitudinal usage study is planned. Practical implications – The implementation of the LL structure was well-received in one government user site and other industry user sites are pending. Practitioners validated the design logic for improving capture and reuse of lessons to render themeasily translatable to a new learner’s context. Originality/value – The new LL structure is uniquely grounded in user needs, developed from existing best practice and is an original application of normative and situation theory to provide consistent rule logic for context/content structure

    Towards an Intelligent Workflow Designer based on the Reuse of Workflow Patterns

    Get PDF
    In order to perform process-aware information systems we need sophisticated methods and concepts for designing and modeling processes. Recently, research on workflow patterns has emerged in order to increase the reuse of recurring workflow structures. However, current workflow modeling tools do not provide functionalities that enable users to define, query, and reuse workflow patterns properly. In this paper we gather a suite for both process modeling and normalization based on workflow patterns reuse. This suite must be used in the extension of some workflow design tool. The suite comprises components for the design of processes from both legacy systems and process modeling

    Radio frequency optimization of a Global System for Mobile (GSM) network

    Get PDF
    Includes bibliographical references
    • …
    corecore