39,261 research outputs found
Two-Stage Multi-Objective Meta-Heuristics for Environmental and Cost-Optimal Energy Refurbishment at District Level
Energy efficiency and environmental performance optimization at the district level are following an upward trend mostly triggered by minimizing the Global Warming Potential (GWP) to 20% by 2020 and 40% by 2030 settled by the European Union (EU) compared with 1990 levels. This paper advances over the state of the art by proposing two novel multi-objective algorithms, named Non-dominated Sorting Genetic Algorithm (NSGA-II) and Multi-Objective Harmony Search (MOHS), aimed at achieving cost-effective energy refurbishment scenarios and allowing at district level the decision-making procedure. This challenge is not trivial since the optimisation process must provide feasible solutions for a simultaneous environmental and economic assessment at district scale taking into consideration highly demanding real-based constraints regarding district and buildings’ specific requirements. Consequently, in this paper, a two-stage optimization methodology is proposed in order to reduce the energy demand and fossil fuel consumption with an affordable investment cost at building level and minimize the total payback time while minimizing the GWP at district level. Aimed at demonstrating the effectiveness of the proposed two-stage multi-objective approaches, this work presents simulation results at two real district case studies in Donostia-San Sebastian (Spain) for which up to a 30% of reduction of GWP at district level is obtained for a Payback Time (PT) of 2–3 years.Part of this work has been developed from results obtained during the H2020 “Optimised Energy
Efficient Design Platform for Refurbishment at District Level” (OptEEmAL) project, Grant No. 680676
Recommended from our members
The road to fully integrated DC-DC conversion via the switched-capacitor approach
This paper provides a perspective on progress toward realization of efficient, fully integrated dc-dc conversion and regulation functionality in CMOS platforms. In providing a comparative assessment between the inductor-based and switched-capacitor approaches, the presentation reviews the salient features in effectiveness in utilization of switch technology and in use and implementation of passives. The analytical conclusions point toward the strong advantages of the switched-capacitor (SC) approach with respect to both switch utilization and much higher energy densities of capacitors versus inductors. The analysis is substantiated with a review of recently developed and published integrated dc-dc converters of both the inductor-based and SC types. © 2012 IEEE
Engineering Crowdsourced Stream Processing Systems
A crowdsourced stream processing system (CSP) is a system that incorporates
crowdsourced tasks in the processing of a data stream. This can be seen as
enabling crowdsourcing work to be applied on a sample of large-scale data at
high speed, or equivalently, enabling stream processing to employ human
intelligence. It also leads to a substantial expansion of the capabilities of
data processing systems. Engineering a CSP system requires the combination of
human and machine computation elements. From a general systems theory
perspective, this means taking into account inherited as well as emerging
properties from both these elements. In this paper, we position CSP systems
within a broader taxonomy, outline a series of design principles and evaluation
metrics, present an extensible framework for their design, and describe several
design patterns. We showcase the capabilities of CSP systems by performing a
case study that applies our proposed framework to the design and analysis of a
real system (AIDR) that classifies social media messages during time-critical
crisis events. Results show that compared to a pure stream processing system,
AIDR can achieve a higher data classification accuracy, while compared to a
pure crowdsourcing solution, the system makes better use of human workers by
requiring much less manual work effort
Active Sampling-based Binary Verification of Dynamical Systems
Nonlinear, adaptive, or otherwise complex control techniques are increasingly
relied upon to ensure the safety of systems operating in uncertain
environments. However, the nonlinearity of the resulting closed-loop system
complicates verification that the system does in fact satisfy those
requirements at all possible operating conditions. While analytical proof-based
techniques and finite abstractions can be used to provably verify the
closed-loop system's response at different operating conditions, they often
produce conservative approximations due to restrictive assumptions and are
difficult to construct in many applications. In contrast, popular statistical
verification techniques relax the restrictions and instead rely upon
simulations to construct statistical or probabilistic guarantees. This work
presents a data-driven statistical verification procedure that instead
constructs statistical learning models from simulated training data to separate
the set of possible perturbations into "safe" and "unsafe" subsets. Binary
evaluations of closed-loop system requirement satisfaction at various
realizations of the uncertainties are obtained through temporal logic
robustness metrics, which are then used to construct predictive models of
requirement satisfaction over the full set of possible uncertainties. As the
accuracy of these predictive statistical models is inherently coupled to the
quality of the training data, an active learning algorithm selects additional
sample points in order to maximize the expected change in the data-driven model
and thus, indirectly, minimize the prediction error. Various case studies
demonstrate the closed-loop verification procedure and highlight improvements
in prediction error over both existing analytical and statistical verification
techniques.Comment: 23 page
From M-ary Query to Bit Query: a new strategy for efficient large-scale RFID identification
The tag collision avoidance has been viewed as one of the most important research problems in RFID communications and bit tracking technology has been widely embedded in query tree (QT) based algorithms to tackle such challenge. Existing solutions show further opportunity to greatly improve the reading performance because collision queries and empty queries are not fully explored. In this paper, a bit query (BQ) strategy based Mary query tree protocol (BQMT) is presented, which can not only eliminate idle queries but also separate collided tags into many small subsets and make full use of the collided bits. To further optimize the reading performance, a modified dual prefixes matching (MDPM) mechanism is presented to allow multiple tags to respond in the same slot and thus significantly reduce the number of queries. Theoretical analysis and simulations are supplemented to validate the effectiveness of the proposed BQMT and MDPM, which outperform the existing QT-based algorithms. Also, the BQMT and MDPM can be combined to BQMDPM to improve the reading performance in system efficiency, total identification time, communication complexity and average energy cost
- …