29,007 research outputs found

    Cognitive apprenticeship : teaching the craft of reading, writing, and mathtematics

    Get PDF
    Includes bibliographical references (p. 25-27)This research was supported by the National Institute of Education under Contract no. US-NIE-C-400-81-0030 and the Office of Naval Research under Contract No. N00014-85-C-002

    The assessment of usability of electronic shopping: A heuristic evaluation

    Get PDF
    Today there are thousands of electronic shops accessible via the Web. Some provide user-friendly features whilst others seem not to consider usability factors at all. Yet, it is critical that the electronic shopping interface is user-friendly so as to help users to obtain their desired results. This study applied heuristic evaluation to examine the usability of current electronic shopping. In particular, it focused on four UK-based supermarkets offering electronic services: including ASDA, Iceland, Sainsbury, and Tesco. The evaluation consists of two stages: a free-flow inspection and a task-based inspection. The results indicate that the most significant and common usability problems have been found to lie within the areas of ‘User Control and Freedom’ and ‘Help and Documentation’. The findings of this study are applied to develop a set of usability guidelines to support the future design of effective interfaces for electronic shopping

    Designing IS service strategy: an information acceleration approach

    Get PDF
    Information technology-based innovation involves considerable risk that requires insight and foresight. Yet, our understanding of how managers develop the insight to support new breakthrough applications is limited and remains obscured by high levels of technical and market uncertainty. This paper applies a new experimental method based on “discrete choice analysis” and “information acceleration” to directly examine how decisions are made in a way that is behaviourally sound. The method is highly applicable to information systems researchers because it provides relative importance measures on a common scale, greater control over alternate explanations and stronger evidence of causality. The practical implications are that information acceleration reduces the levels of uncertainty and generates a more accurate rationale for IS service strategy decisions

    Layered evaluation of interactive adaptive systems : framework and formative methods

    Get PDF
    Peer reviewedPostprin

    Automated generation of computationally hard feature models using evolutionary algorithms

    Get PDF
    This is the post-print version of the final paper published in Expert Systems with Applications. The published article is available from the link below. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. Copyright @ 2014 Elsevier B.V.A feature model is a compact representation of the products of a software product line. The automated extraction of information from feature models is a thriving topic involving numerous analysis operations, techniques and tools. Performance evaluations in this domain mainly rely on the use of random feature models. However, these only provide a rough idea of the behaviour of the tools with average problems and are not sufficient to reveal their real strengths and weaknesses. In this article, we propose to model the problem of finding computationally hard feature models as an optimization problem and we solve it using a novel evolutionary algorithm for optimized feature models (ETHOM). Given a tool and an analysis operation, ETHOM generates input models of a predefined size maximizing aspects such as the execution time or the memory consumption of the tool when performing the operation over the model. This allows users and developers to know the performance of tools in pessimistic cases providing a better idea of their real power and revealing performance bugs. Experiments using ETHOM on a number of analyses and tools have successfully identified models producing much longer executions times and higher memory consumption than those obtained with random models of identical or even larger size.European Commission (FEDER), the Spanish Government and the Andalusian Government

    Comparative study of different approaches to solve batch process scheduling and optimisation problems

    Get PDF
    Effective approaches are important to batch process scheduling problems, especially those with complex constraints. However, most research focus on improving optimisation techniques, and those concentrate on comparing their difference are inadequate. This study develops an optimisation model of batch process scheduling problems with complex constraints and investigates the performance of different optimisation techniques, such as Genetic Algorithm (GA) and Constraint Programming (CP). It finds that CP has a better capacity to handle batch process problems with complex constraints but it costs longer time

    Analysing Process Models Quantitatively

    Get PDF
    Over the years, there has been much interest in modelling processes. Processes include those associated with the development of software and those business processes that make use of software systems. Recent research in Systems Engineering for Business Process Change highlights the importance of modelling business processes in order to evolve and maintain the legacy systems that support those processes. Business processes are typically described with static (diagrammatic) models. This paper illustrates how quantitative techniques can facilitate analysis of such models. This is illustrated with reference to the process modelling notation Role Activity Diagrams (RADs). An example process, taken from an investigation of the bidding process of a large telecommunications systems supplier, is used to show how a quantitative approach can be used to highlight features in RADs that are useful to the process modeller. We show how simple measures reveal high levels of role coupling and discrepancies between different perspectives. Since the models are non-trivial — there are 101 roles and almost 300 activities — we argue that quantitative analysis can be a useful adjunct for the modeller

    Two-echelon freight transport optimisation: unifying concepts via a systematic review

    Get PDF
    Multi-echelon distribution schemes are one of the most common strategies adopted by the transport companies in an aim of cost reduction, but their identification in scientific literature is not always easy due to a lack of unification. This paper presents the main concepts of two-echelon distribution via a systematic review, in the specific a meta-narrative analysis, in order to identify and unify the main concepts, issues and methods that can be helpful for scientists and transport practitioners. The problem of system cost optimisation in two-echelon freight transport systems is defined. Moreover, the main variants are synthetically presented and discussed. Finally, future research directions are proposed.location-routing problems, multi-echelon distribution, cross-docking, combinatorial optimisation, systematic review.
    • 

    corecore