43,399 research outputs found

    Death and Paperwork Reduction

    Get PDF
    How does government value people\u27s time? Often the valuation is implicit, even mysterious. But in patches of the federal administrative state, paperwork burdens are quantified in hours and often monetized. When agencies do monetize, they look to how the labor market values the time of the people faced with paperwork. The result is that some people\u27s time is valued over ten times more than other people\u27s time. In contrast, when agencies monetize the value of statistical life for cost-benefit analysis, they look to how people faced with a risk of death subjectively value its reduction. In practice, agencies assign the same value to every statistical life saved by a given policy. This Article establishes these patterns of agency behavior and suggests that there is no satisfying justification for them. Welfarist and egalitarian principles, along with the logic of statistical life valuation, lean against the use of market wages to monetize a person\u27s time doing government paperwork. The impact of this practice might be limited, given the modest ambition of today\u27s paperwork reduction efforts. But time-related burdens—and benefits—are key consequences of government decisions in countless contexts. If we want to scale up a thoughtful process for valuing people\u27s time in the future, we will need new foundations

    A Programming Environment Evaluation Methodology for Object-Oriented Systems

    Get PDF
    The object-oriented design strategy as both a problem decomposition and system development paradigm has made impressive inroads into the various areas of the computing sciences. Substantial development productivity improvements have been demonstrated in areas ranging from artificial intelligence to user interface design. However, there has been very little progress in the formal characterization of these productivity improvements and in the identification of the underlying cognitive mechanisms. The development and validation of models and metrics of this sort require large amounts of systematically-gathered structural and productivity data. There has, however, been a notable lack of systematically-gathered information on these development environments. A large part of this problem is attributable to the lack of a systematic programming environment evaluation methodology that is appropriate to the evaluation of object-oriented systems

    Five year review of standards: GCSE mathematics

    Get PDF

    Review of research in feature-based design

    Get PDF
    Research in feature-based design is reviewed. Feature-based design is regarded as a key factor towards CAD/CAPP integration from a process planning point of view. From a design point of view, feature-based design offers possibilities for supporting the design process better than current CAD systems do. The evolution of feature definitions is briefly discussed. Features and their role in the design process and as representatives of design-objects and design-object knowledge are discussed. The main research issues related to feature-based design are outlined. These are: feature representation, features and tolerances, feature validation, multiple viewpoints towards features, features and standardization, and features and languages. An overview of some academic feature-based design systems is provided. Future research issues in feature-based design are outlined. The conclusion is that feature-based design is still in its infancy, and that more research is needed for a better support of the design process and better integration with manufacturing, although major advances have already been made

    Lost in translation: Exposing hidden compiler optimization opportunities

    Get PDF
    Existing iterative compilation and machine-learning-based optimization techniques have been proven very successful in achieving better optimizations than the standard optimization levels of a compiler. However, they were not engineered to support the tuning of a compiler's optimizer as part of the compiler's daily development cycle. In this paper, we first establish the required properties which a technique must exhibit to enable such tuning. We then introduce an enhancement to the classic nightly routine testing of compilers which exhibits all the required properties, and thus, is capable of driving the improvement and tuning of the compiler's common optimizer. This is achieved by leveraging resource usage and compilation information collected while systematically exploiting prefixes of the transformations applied at standard optimization levels. Experimental evaluation using the LLVM v6.0.1 compiler demonstrated that the new approach was able to reveal hidden cross-architecture and architecture-dependent potential optimizations on two popular processors: the Intel i5-6300U and the Arm Cortex-A53-based Broadcom BCM2837 used in the Raspberry Pi 3B+. As a case study, we demonstrate how the insights from our approach enabled us to identify and remove a significant shortcoming of the CFG simplification pass of the LLVM v6.0.1 compiler.Comment: 31 pages, 7 figures, 2 table. arXiv admin note: text overlap with arXiv:1802.0984
    • …
    corecore