1 research outputs found

    Evaluation of Domain-specific Rule Generation Framework based on Usability Criteria

    Get PDF
    The challenges faced by domain experts, commitments made to domain-specific rule (DSR) languages and process design are described. We investigate the business application developments and existing challenges of evaluation strategies of DSR articulations. Often, multiple domain scenarios pose end-user predicaments complicating the computational ability of DSR. In addition, implementation of DSR and its configuration are belated due to poorly evaluated usability criteria. A new framework is needed, facilitating the DSR language and enhancing the computational intelligence. We intend to evaluate the performance of DSR generation and framework integration with variety of usability conditions including efficiency and effectiveness of configuration through system usability score (SUS). Empirical research involving experimental data, questionnaire surveys, and interview outcomes provide conclusive evaluation attributes and their fact instances from SUS. Both manual and semi-automatic configurations are tested. Semiautomatic configuration appears to be more efficient and satisfactory with regard to artifact performance, quality, learnability, user-friendly and reliability
    corecore