39,261 research outputs found

    Two-Stage Multi-Objective Meta-Heuristics for Environmental and Cost-Optimal Energy Refurbishment at District Level

    Get PDF
    Energy efficiency and environmental performance optimization at the district level are following an upward trend mostly triggered by minimizing the Global Warming Potential (GWP) to 20% by 2020 and 40% by 2030 settled by the European Union (EU) compared with 1990 levels. This paper advances over the state of the art by proposing two novel multi-objective algorithms, named Non-dominated Sorting Genetic Algorithm (NSGA-II) and Multi-Objective Harmony Search (MOHS), aimed at achieving cost-effective energy refurbishment scenarios and allowing at district level the decision-making procedure. This challenge is not trivial since the optimisation process must provide feasible solutions for a simultaneous environmental and economic assessment at district scale taking into consideration highly demanding real-based constraints regarding district and buildings’ specific requirements. Consequently, in this paper, a two-stage optimization methodology is proposed in order to reduce the energy demand and fossil fuel consumption with an affordable investment cost at building level and minimize the total payback time while minimizing the GWP at district level. Aimed at demonstrating the effectiveness of the proposed two-stage multi-objective approaches, this work presents simulation results at two real district case studies in Donostia-San Sebastian (Spain) for which up to a 30% of reduction of GWP at district level is obtained for a Payback Time (PT) of 2–3 years.Part of this work has been developed from results obtained during the H2020 “Optimised Energy Efficient Design Platform for Refurbishment at District Level” (OptEEmAL) project, Grant No. 680676

    Engineering Crowdsourced Stream Processing Systems

    Full text link
    A crowdsourced stream processing system (CSP) is a system that incorporates crowdsourced tasks in the processing of a data stream. This can be seen as enabling crowdsourcing work to be applied on a sample of large-scale data at high speed, or equivalently, enabling stream processing to employ human intelligence. It also leads to a substantial expansion of the capabilities of data processing systems. Engineering a CSP system requires the combination of human and machine computation elements. From a general systems theory perspective, this means taking into account inherited as well as emerging properties from both these elements. In this paper, we position CSP systems within a broader taxonomy, outline a series of design principles and evaluation metrics, present an extensible framework for their design, and describe several design patterns. We showcase the capabilities of CSP systems by performing a case study that applies our proposed framework to the design and analysis of a real system (AIDR) that classifies social media messages during time-critical crisis events. Results show that compared to a pure stream processing system, AIDR can achieve a higher data classification accuracy, while compared to a pure crowdsourcing solution, the system makes better use of human workers by requiring much less manual work effort

    Active Sampling-based Binary Verification of Dynamical Systems

    Full text link
    Nonlinear, adaptive, or otherwise complex control techniques are increasingly relied upon to ensure the safety of systems operating in uncertain environments. However, the nonlinearity of the resulting closed-loop system complicates verification that the system does in fact satisfy those requirements at all possible operating conditions. While analytical proof-based techniques and finite abstractions can be used to provably verify the closed-loop system's response at different operating conditions, they often produce conservative approximations due to restrictive assumptions and are difficult to construct in many applications. In contrast, popular statistical verification techniques relax the restrictions and instead rely upon simulations to construct statistical or probabilistic guarantees. This work presents a data-driven statistical verification procedure that instead constructs statistical learning models from simulated training data to separate the set of possible perturbations into "safe" and "unsafe" subsets. Binary evaluations of closed-loop system requirement satisfaction at various realizations of the uncertainties are obtained through temporal logic robustness metrics, which are then used to construct predictive models of requirement satisfaction over the full set of possible uncertainties. As the accuracy of these predictive statistical models is inherently coupled to the quality of the training data, an active learning algorithm selects additional sample points in order to maximize the expected change in the data-driven model and thus, indirectly, minimize the prediction error. Various case studies demonstrate the closed-loop verification procedure and highlight improvements in prediction error over both existing analytical and statistical verification techniques.Comment: 23 page

    From M-ary Query to Bit Query: a new strategy for efficient large-scale RFID identification

    Get PDF
    The tag collision avoidance has been viewed as one of the most important research problems in RFID communications and bit tracking technology has been widely embedded in query tree (QT) based algorithms to tackle such challenge. Existing solutions show further opportunity to greatly improve the reading performance because collision queries and empty queries are not fully explored. In this paper, a bit query (BQ) strategy based Mary query tree protocol (BQMT) is presented, which can not only eliminate idle queries but also separate collided tags into many small subsets and make full use of the collided bits. To further optimize the reading performance, a modified dual prefixes matching (MDPM) mechanism is presented to allow multiple tags to respond in the same slot and thus significantly reduce the number of queries. Theoretical analysis and simulations are supplemented to validate the effectiveness of the proposed BQMT and MDPM, which outperform the existing QT-based algorithms. Also, the BQMT and MDPM can be combined to BQMDPM to improve the reading performance in system efficiency, total identification time, communication complexity and average energy cost
    • …
    corecore