284,129 research outputs found

    A review of daylighting design and implementation in buildings

    Get PDF

    An integrated shipment planning and storage capacity decision under uncertainty: a simulation study

    Get PDF
    Purpose – In transportation and distribution systems, the shipment decisions, fleet capacity, and storage capacity are interrelated in a complex way, especially when the authors take into account uncertainty of the demand rate and shipment lead time. While shipment planning is tactical or operational in nature, increasing storage capacity often requires top management’s authority. The purpose of this paper is to present a new method to integrate both operational and strategic decision parameters, namely shipment planning and storage capacity decision under uncertainty. The ultimate goal is to provide a near optimal solution that leads to a striking balance between the total logistics costs and product availability, critical in maritime logistics of bulk shipment of commodity items. Design/methodology/approach – The authors use simulation as research method. The authors develop a simulation model to investigate the effects of various factors on costs and service levels of a distribution system. The model mimics the transportation and distribution problems of bulk cement in a major cement company in Indonesia consisting of a silo at the port of origin, two silos at two ports of destination, and a number of ships that transport the bulk cement. The authors develop a number of “what-if” scenarios by varying the storage capacity at the port of origin as well as at the ports of destinations, number of ships operated, operating hours of ports, and dispatching rules for the ships. Each scenario is evaluated in terms of costs and service level. A full factorial experiment has been conducted and analysis of variance has been used to analyze the results. Findings – The results suggest that the number of ships deployed, silo capacity, working hours of ports, and the dispatching rules of ships significantly affect both total costs and service level. Interestingly, operating fewer ships enables the company to achieve almost the same service level and gaining substantial cost savings if constraints in other part of the system are alleviated, i.e., storage capacities and working hours of ports are extended. Practical implications – Cost is a competitive factor for bulk items like cement, and thus the proposed scenarios could be implemented by the company to substantially reduce the transportation and distribution costs. Alleviating storage capacity constraint is obviously an idea that needs to be considered when optimizing shipment planning alone could not give significant improvements. Originality/value – Existing research has so far focussed on the optimization of shipment planning/scheduling, and considers shipment planning/scheduling as the objective function while treating the storage capacity as constraints. The simulation model enables “what-if” analyses to be performed and has overcome the difficulties and impracticalities of analytical methods especially when the system incorporates stochastic variables exhibited in the case example. The use of efficient frontier analysis for analyzing the simulation results is a novel idea which has been proven to be effective in screening non-dominated solutions. This has provided the authors with near optimal solutions to trade-off logistics costs and service levels (availability), with minimal experimentation times

    Data-driven modelling of biological multi-scale processes

    Full text link
    Biological processes involve a variety of spatial and temporal scales. A holistic understanding of many biological processes therefore requires multi-scale models which capture the relevant properties on all these scales. In this manuscript we review mathematical modelling approaches used to describe the individual spatial scales and how they are integrated into holistic models. We discuss the relation between spatial and temporal scales and the implication of that on multi-scale modelling. Based upon this overview over state-of-the-art modelling approaches, we formulate key challenges in mathematical and computational modelling of biological multi-scale and multi-physics processes. In particular, we considered the availability of analysis tools for multi-scale models and model-based multi-scale data integration. We provide a compact review of methods for model-based data integration and model-based hypothesis testing. Furthermore, novel approaches and recent trends are discussed, including computation time reduction using reduced order and surrogate models, which contribute to the solution of inference problems. We conclude the manuscript by providing a few ideas for the development of tailored multi-scale inference methods.Comment: This manuscript will appear in the Journal of Coupled Systems and Multiscale Dynamics (American Scientific Publishers

    Molecular modeling to study dendrimers for biomedical applications

    Get PDF
    © 2014 by the authors; licensee MDPI; Basel; Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/4.0/). Date of Acceptance: 17/11/2014Molecular modeling techniques provide a powerful tool to study the properties of molecules and their interactions at the molecular level. The use of computational techniques to predict interaction patterns and molecular properties can inform the design of drug delivery systems and therapeutic agents. Dendrimers are hyperbranched macromolecular structures that comprise repetitive building blocks and have defined architecture and functionality. Their unique structural features can be exploited to design novel carriers for both therapeutic and diagnostic agents. Many studies have been performed to iteratively optimise the properties of dendrimers in solution as well as their interaction with drugs, nucleic acids, proteins and lipid membranes. Key features including dendrimer size and surface have been revealed that can be modified to increase their performance as drug carriers. Computational studies have supported experimental work by providing valuable insights about dendrimer structure and possible molecular interactions at the molecular level. The progress in computational simulation techniques and models provides a basis to improve our ability to better predict and understand the biological activities and interactions of dendrimers. This review will focus on the use of molecular modeling tools for the study and design of dendrimers, with particular emphasis on the efforts that have been made to improve the efficacy of this class of molecules in biomedical applications.Peer reviewedFinal Published versio

    Model-based dependability analysis : state-of-the-art, challenges and future outlook

    Get PDF
    Abstract: Over the past two decades, the study of model-based dependability analysis has gathered significant research interest. Different approaches have been developed to automate and address various limitations of classical dependability techniques to contend with the increasing complexity and challenges of modern safety-critical system. Two leading paradigms have emerged, one which constructs predictive system failure models from component failure models compositionally using the topology of the system. The other utilizes design models - typically state automata - to explore system behaviour through fault injection. This paper reviews a number of prominent techniques under these two paradigms, and provides an insight into their working mechanism, applicability, strengths and challenges, as well as recent developments within these fields. We also discuss the emerging trends on integrated approaches and advanced analysis capabilities. Lastly, we outline the future outlook for model-based dependability analysis

    ASCR/HEP Exascale Requirements Review Report

    Full text link
    This draft report summarizes and details the findings, results, and recommendations derived from the ASCR/HEP Exascale Requirements Review meeting held in June, 2015. The main conclusions are as follows. 1) Larger, more capable computing and data facilities are needed to support HEP science goals in all three frontiers: Energy, Intensity, and Cosmic. The expected scale of the demand at the 2025 timescale is at least two orders of magnitude -- and in some cases greater -- than that available currently. 2) The growth rate of data produced by simulations is overwhelming the current ability, of both facilities and researchers, to store and analyze it. Additional resources and new techniques for data analysis are urgently needed. 3) Data rates and volumes from HEP experimental facilities are also straining the ability to store and analyze large and complex data volumes. Appropriately configured leadership-class facilities can play a transformational role in enabling scientific discovery from these datasets. 4) A close integration of HPC simulation and data analysis will aid greatly in interpreting results from HEP experiments. Such an integration will minimize data movement and facilitate interdependent workflows. 5) Long-range planning between HEP and ASCR will be required to meet HEP's research needs. To best use ASCR HPC resources the experimental HEP program needs a) an established long-term plan for access to ASCR computational and data resources, b) an ability to map workflows onto HPC resources, c) the ability for ASCR facilities to accommodate workflows run by collaborations that can have thousands of individual members, d) to transition codes to the next-generation HPC platforms that will be available at ASCR facilities, e) to build up and train a workforce capable of developing and using simulations and analysis to support HEP scientific research on next-generation systems.Comment: 77 pages, 13 Figures; draft report, subject to further revisio
    corecore