1,290 research outputs found

    Wilt Chamberlain Redux: Thinking Clearly about Externalities and the Promises of Justice

    Get PDF
    Gordon Barnes accuses Robert Nozick and Eric Mack of neglecting, in two ways, the practical, empirical questions relevant to justice in the real world.1 He thinks these omissions show that the argument behind the Wilt Chamberlain example—which Nozick famously made in his seminal Anarchy, State, and Utopia—fails. As a result, he suggests that libertarians should concede that this argument fails. In this article, we show that Barnes’s key arguments hinge on misunderstandings of, or failures to notice, key aspects of the entitlement theory that undergirds Nozick’s and Mack’s work. Once the theory is properly understood, Barnes’s challenges fail to undermine the Chamberlain example, in particular, and the entitlement theory, in general

    Restaurant Design: An Industrial Engineering Perspective

    Get PDF
    The objective of this paper is to design a health food restaurant using current industrial engineering tools and practices. This includes creating a menu, determining the average gross profit per customer, determining the location, designing a 3-d layout, simulating service, and analyzing local food supply chains. Using a simplified menu and pricing quotes from a food distributor, the average gross profit per customer is estimated to be 3.50.Locationisdeterminedbycomparingthepredictednumberofcustomersandthecosttorentattwodifferentareas.Adowntownlocationisdeterminedtogrossapproximately3.50. Location is determined by comparing the predicted number of customers and the cost to rent at two different areas. A downtown location is determined to gross approximately 5,200 per month over a residential location in addition to having increased exposure. The 3-d layout promotes flow of customers while reducing the amount of distance employees must travel by placing items and food storage where they are needed. The simulation predicted that 2 servers will be optimal for 20 to 40 customers per hour whereas 3 servers will be optimal for 40 to 60 customers per hour. Using a local supply will be ideal only when supplemented with a food distributor. The viability of this restaurant is more dependent on business principles and is outside the scope of the project

    Estimation of the flammability zone boundaries with thermodynamic and empirical equations

    Get PDF
    The flammability zone boundaries are very important properties to prevent explosions in the process industries. Within the boundaries, a flame or explosion can occur so it is important to understand these boundaries to prevent fires and explosions. Very little work has been reported in the literature to model the flammability zone boundaries. Two boundaries are defined and studied: the upper flammability zone boundary and the lower flammability zone boundary. Three methods are presented to predict the upper and lower flammability zone boundaries: The linear model The extended linear model, and An empirical model The linear model is a thermodynamic model that uses the upper flammability limit (UFL) and lower flammability limit (LFL) to calculate two adiabatic flame temperatures. When the proper assumptions are applied, the linear model can be reduced to the well-known equation yLOC = zyLFL for estimation of the limiting oxygen concentration. The extended linear model attempts to account for the changes in the reactions along the UFL boundary. Finally, the empirical method fits the boundaries with linear equations between the UFL or LFL and the intercept with the oxygen axis. xx Comparison of the models to experimental data of the flammability zone shows that the best model for estimating the flammability zone boundaries is the empirical method. It is shown that is fits the limiting oxygen concentration (LOC), upper oxygen limit (UOL), and the lower oxygen limit (LOL) quite well. The regression coefficient values for the fits to the LOC, UOL, and LOL are 0.672, 0.968, and 0.959, respectively. This is better than the fit of the zyLFL method for the LOC in which the regression coefficient’s value is 0.416

    Plume Development of the Shoemaker-Levy 9 Comet Impact

    Get PDF
    We have studied plume formation after a Jovian comet impact using the ZEUS-MP 2 hydrodynamics code. The three-dimensional models followed objects with 500, 750, and 1000 meter diameters. Our simulations show the development of a fast, upward-moving component of the plume in the wake of the impacting comet that "pinches off" from the bulk of the cometary material ~50 km below the 1 bar pressure level, ~100 km above the depth of greatest mass and energy deposition. The fast-moving component contains about twice the mass of the initial comet, but consists almost entirely (>99.9%) of Jovian atmosphere rather than cometary material. The ejecta rise mainly along the impact trajectory, but an additional vertical velocity component due to buoyancy establishes itself within seconds of impact, leading to an asymmetry in the ejecta with respect to the entry trajectory. The mass of the upward-moving component follows a velocity distribution M(>v) approximately proportional to v^-1.4 (v^-1.6 for the 750 m and 500 m cases) in the velocity range 0.1 < v < 10 km/s.Comment: 5 pages, 4 figures. Accepted for publication in The Astrophysical Journa

    Carbon Monoxide Packaging for Fresh Pork

    Get PDF
    Injected and non-injected pork chops were packaged in aerobic, vacuum and modified atmosphere [with and without carbon monoxide (CO)] packages to evaluate quality and shelf life during refrigerated storage.Color stability and attractiveness was dramatically improved by the carbon monoxide treatment over any other packaging treatment in this study.This effect permits use of high carbon dioxide atmosphere for packaging of fresh meat to achieve improved shelf life for both color and microbial qualit

    Connections Between Adaptive Control and Optimization in Machine Learning

    Full text link
    This paper demonstrates many immediate connections between adaptive control and optimization methods commonly employed in machine learning. Starting from common output error formulations, similarities in update law modifications are examined. Concepts in stability, performance, and learning, common to both fields are then discussed. Building on the similarities in update laws and common concepts, new intersections and opportunities for improved algorithm analysis are provided. In particular, a specific problem related to higher order learning is solved through insights obtained from these intersections.Comment: 18 page

    Processing Color in Astronomical Imagery

    Get PDF
    Every year, hundreds of images from telescopes on the ground and in space are released to the public, making their way into popular culture through everything from computer screens to postage stamps. These images span the entire electromagnetic spectrum from radio waves to infrared light to X-rays and gamma rays, a majority of which is undetectable to the human eye without technology. Once these data are collected, one or more specialists must process the data to create an image. Therefore, the creation of astronomical imagery involves a series of choices. How do these choices affect the comprehension of the science behind the images? What is the best way to represent data to a non-expert? Should these choices be based on aesthetics, scientific veracity, or is it possible to satisfy both? This paper reviews just one choice out of the many made by astronomical image processors: color. The choice of color is one of the most fundamental when creating an image taken with modern telescopes. We briefly explore the concept of the image as translation, particularly in the case of astronomical images from invisible portions of the electromagnetic spectrum. After placing modern astronomical imagery and photography in general in the context of its historical beginnings, we review the standards (or lack thereof) in making the basic choice of color. We discuss the possible implications for selecting one color palette over another in the context of the appropriateness of using these images as science communication products with a specific focus on how the non-expert perceives these images and how that affects their trust in science. Finally, we share new data sets that begin to look at these issues in scholarly research and discuss the need for a more robust examination of this and other related topics in the future to better understand the implications for science communications.Comment: 10 pages, 6 figures, published in Studies in Media and Communicatio

    Multimodal infrastructure Investment Decision Making: An Institutional and Funding Perspective

    Get PDF
    This paper presents a case analysis of multimodalism in transportation investment decision making as it relates to other policy decisions, including infrastructure ownership structures, geographic scales, and revenue sources. This allows a more complete understanding of multimodalism’s benefits, drawbacks, and opportunities. Private infrastructure developers are more likely to evaluate and select investments in a multimodal fashion than either the public sector or public-private partnerships. Decentralized, municipal decision making tends to reduce barriers to multimodalism, although with reduced capacity for large-scale project implementation. Experience suggests that the source of transportation revenues is a less important influence on multimodalism than the organizations collecting them

    Critical Opportunity Areas for Building Performance

    Get PDF
    Green buildings are a proposed holistic solution to reduce energy consumption while simultaneously improving an array of factors affecting the indoor quality of life for building occupants. However, green building performance varies and may not achieve intended design goals. Research has concluded that no single factor determines the actual energy performance of buildings. To deliver energy-efficient buildings an integrated design that considers climate, technology, operation and maintenance and occupant behavior should be implemented. This work aimed to employ a holistic lens to relate human-building interaction and building performance characteristics. Specifically, systems theory and complex-problem solving techniques were employed to capture the dynamic interactions between the social and technical parts and processes of building systems and identify gaps causing the underperformance of buildings. Synergies not captured in the current design process but impact the ability of a building system to achieve its design goals were outlined. Performance metrics that a single system inadvertently affects along social, physical and economic dimensions were identified as well as high-impact opportunity areas for the creation of high-performance buildings. Addressing these synergies in the building equipment and full building design will enable stakeholder-centered systems integration, improving the efficiency and efficacy of buildings
    • …
    corecore