2,011 research outputs found
Cold Chain Energy Analysis for Sustainable Food and Beverage Supply
Perishable goods, such as chilled and frozen foods, have a short shelf life and high sensitivity to their surrounding environment (e.g., temperature, humidity, and light intensity). For this reason, they must be distributed within a specific time and require special equipment and facilities (e.g., refrigeration and dehumidification systems) throughout the entire chain from farm to fork to ensure slow deterioration and to deliver safe and high-quality products to consumers. Cold chains can last for short periods, such as a few hours, or for several months or even years (e.g., frozen food products) depending on the product and the target market. A huge amount of energy is required to preserve quality by maintaining the desired temperature level over time. The required energy is also affected by inventory management policies (e.g., warehouse filling levels affect the cooling demand per unit of a product) and the behavior of the operators (e.g., number and duration of door openings). Furthermore, waste entails the loss of energy and other resources consumed for processing and storing these products. The aim of the present study is to propose a quantitative approach in order to map the energy flows throughout the cold chain in the food and beverage sector and to evaluate the overall energy performance. The results of the energy flow mapping give decisionmakers insights into the minimum energy required by the cold chain and allow them to prioritize energy efficiency measures by detecting the most energy consuming stages of the cold chain. The implementation of a holistic approach, shifting from a single-company perspective to chain assessment, leads to a global optimum and to an increased implementation rate of energy efficiency measures due to the reduced barriers perceived by different actors of the cold chain
Minimax Classification with 0-1 Loss and Performance Guarantees
Supervised classification techniques use training samples to find classification
rules with small expected 0-1 loss. Conventional methods achieve efficient learning
and out-of-sample generalization by minimizing surrogate losses over specific
families of rules. This paper presents minimax risk classifiers (MRCs) that do not
rely on a choice of surrogate loss and family of rules. MRCs achieve efficient
learning and out-of-sample generalization by minimizing worst-case expected 0-1
loss w.r.t. uncertainty sets that are defined by linear constraints and include the
true underlying distribution. In addition, MRCsâ learning stage provides performance
guarantees as lower and upper tight bounds for expected 0-1 loss. We also
present MRCsâ finite-sample generalization bounds in terms of training size and
smallest minimax risk, and show their competitive classification performance w.r.t.
state-of-the-art techniques using benchmark datasets.Ramon y Cajal Grant RYC-2016-1938
Assessment of Energy Efficiency Measures in Food Cold Supply Chains: A Dairy Industry Case Study
The quality of human nutrition has acquired significant improvements thanks to the opportunity to store food in suitable temperature conditions. Refrigeration has allowed the slowing of chemical and biological degradation and hence the waste of foodstuff, but at the same time increases energy consumption. These effects impact the environment and the sustainability performance of the cold chain, and drive consumersâ choices. The stakeholders of the chain are, therefore, constantly looking for improvement actions to reduce environmental impacts. This paper aims to provide a methodology for prioritizing and assessing the energy efficiency measures for cold chains in terms of quality losses and specific energy consumption, distinguishing between technological, maintenance, and managerial opportunities. This analysis is based on the cold supply chain tool, developed under the H2020 project ICCEE (âImproving Cold Chain Energy Efficiencyâ) which focuses on a holistic approach, not looking only at the individual stages of the cold chain. Furthermore, an economic evaluation has been proposed considering cost savings and the investment needed
Supply chain finance for ameliorating and deteriorating products: a systematic literature review
Ameliorating and deteriorating products, or, more generally, items that change value over time, present a high sensitiveness to the surrounding environment (e.g., temperature, humidity, and light intensity). For this reason, they should be properly stored along the supply chain to guarantee the desired quality to the consumers. Specifically, ameliorating items face an increase in value if there are stored for longer periods, which can lead to higher selling price. At the same time, the costumersâ demand is sensitive to the price (i.e., the higher the selling price the lower the final demand), sensitiveness that is related to the quality of the products (i.e., lower sensitiveness for high-quality products). On the contrary, deteriorating items lose quality and value over time which result in revenue losses due to lost sales or reduced selling price. Since these products need to be properly stored (i.e., usually in temperature- and humidity-controlled warehouses) the holding costs, which comprise also the energy costs, may be particularly relevant impacting on the economic, environmental, and social sustainability of the supply chain. Furthermore, due to the recent economic crisis, companies (especially, small and medium enterprises) face payment difficulties of customers and high volatility of resources prices. This increases the risk of insolvency and on the other hand the financing needs. In this context, supply chain finance emerged as a mean for efficiency by coordinating the financial flow and providing a set of financial schemes aiming at optimizing accounts payable and receivable along the supply chain. The aim of the present study is thus to investigate through a systematic literature review the two main themes presented (i.e., inventory management models for products that change value over time, and financial techniques and strategies to support companies in inventory management) to understand if any financial technique has been studied for supporting the management of this class of products and to verify the existing literature gap
Searching for dominant high-level features for music information retrieval
Music Information Retrieval systems are often based on the analysis of a large number of low-level audio features. When dealing with problems of musical genre description and visualization, however, it would be desirable to work with a very limited number of highly informative and discriminant macro-descriptors. In this paper we focus on a speciïŹc class of training-based descriptors, which are obtained as the loglikelihood of a Gaussian Mixture Model trained with short musical excerpts that selectively exhibit a certain semantic homogeneity. As these descriptors are critically dependent on the training sets, we approach the problem of how to automatically generate suitable training sets and optimize the associated macro-features in terms of discriminant power and informative impact. We then show the application of a set of three identiïŹed macro-features to genre visualization, tracking and classiïŹcation
Gate stability of GaN-Based HEMTs with P-Type Gate
status: publishe
Detection of Earth-like Planets Using Apodized Telescopes
The mission of NASA's Terrestrial Planet Finder (TPF) is to find Earth-like
planets orbiting other stars and characterize the atmospheres of these planets
using spectroscopy. Because of the enormous brightness ratio between the star
and the reflected light from the planet, techniques must be found to reduce the
brightness of the star. The current favorite approach to doing this is with
interferometry: interfering the light from two or more separated telescopes
with a phase shift, nulling out the starlight. While this technique can,
in principle, achieve the required dynamic range, building a space
interferometer that has the necessary characteristics poses immense technical
difficulties. In this paper, we suggest a much simpler approach to achieving
the required dynamic range. By simply adjusting the transmissive shape of a
telescope aperture, the intensity in large regions around the stellar image can
be reduced nearly to zero. This approach could lead to construction of a TPF
using conventional technologies, requiring space optics on a much smaller scale
than the current TPF approach.Comment: Accepted for publication in ApJ Letters, 9 pages, 6 figure
Frequency-domain nonlinear modeling approaches for power systems components - A comparison
Harmonic simulations play a key role in studying and predicting the impact of nonlinear devices on the power quality level of distribution grids. A frequency-domain approach allows higher computational efficiency, which has key importance as long as complex networks have to be studied. However, this requires proper frequency-domain behavioral models able to represent the nonlinear voltage-current relationship characterizing these devices. The Frequency Transfer Matrix (FTM) method is one of the most widespread frequency domain modeling approaches for power system applications. However, others suitable techniques have been developed in the last years, in particular the X-parameters approach, which comes from radiofrequency and microwave applications, and the simplified Volterra models under quasi-sinusoidal conditions, that have been specifically tailored for power system devices. In this paper FTM, X-parameters and simplified Volterra approaches are compared in representing the nonlinear voltage-current relationship of a bridge rectifier feeding an ohmic-capacitive dc load. Results show that the X-parameters model reaches good accuracy, which is slightly better than that achieved by the FTM and simplified Volterra models, but with a considerably larger set of coefficients. Simplified Volterra models under quasi-sinusoidal conditions allows an effective trade-off between accuracy and complexity
- âŠ