1,721 research outputs found
Recommended from our members
State-of-the-art on research and applications of machine learning in the building life cycle
Fueled by big data, powerful and affordable computing resources, and advanced algorithms, machine learning has been explored and applied to buildings research for the past decades and has demonstrated its potential to enhance building performance. This study systematically surveyed how machine learning has been applied at different stages of building life cycle. By conducting a literature search on the Web of Knowledge platform, we found 9579 papers in this field and selected 153 papers for an in-depth review. The number of published papers is increasing year by year, with a focus on building design, operation, and control. However, no study was found using machine learning in building commissioning. There are successful pilot studies on fault detection and diagnosis of HVAC equipment and systems, load prediction, energy baseline estimate, load shape clustering, occupancy prediction, and learning occupant behaviors and energy use patterns. None of the existing studies were adopted broadly by the building industry, due to common challenges including (1) lack of large scale labeled data to train and validate the model, (2) lack of model transferability, which limits a model trained with one data-rich building to be used in another building with limited data, (3) lack of strong justification of costs and benefits of deploying machine learning, and (4) the performance might not be reliable and robust for the stated goals, as the method might work for some buildings but could not be generalized to others. Findings from the study can inform future machine learning research to improve occupant comfort, energy efficiency, demand flexibility, and resilience of buildings, as well as to inspire young researchers in the field to explore multidisciplinary approaches that integrate building science, computing science, data science, and social science
A Tutorial on Bayesian Optimization of Expensive Cost Functions, with Application to Active User Modeling and Hierarchical Reinforcement Learning
We present a tutorial on Bayesian optimization, a method of finding the
maximum of expensive cost functions. Bayesian optimization employs the Bayesian
technique of setting a prior over the objective function and combining it with
evidence to get a posterior function. This permits a utility-based selection of
the next observation to make on the objective function, which must take into
account both exploration (sampling from areas of high uncertainty) and
exploitation (sampling areas likely to offer improvement over the current best
observation). We also present two detailed extensions of Bayesian optimization,
with experiments---active user modelling with preferences, and hierarchical
reinforcement learning---and a discussion of the pros and cons of Bayesian
optimization based on our experiences
Analyzing Sensor-Based Individual and Population Behavior Patterns via Inverse Reinforcement Learning
Digital markers of behavior can be continuously created, in everyday settings, using time series data collected by ambient sensors. The goal of this work was to perform individual- and population-level behavior analysis from such time series sensor data. In this paper, we introduce a novel algorithm-Resident Relative Entropy-Inverse Reinforcement Learning (RRE-IRL)-to perform an analysis of a single smart home resident or a group of residents, using inverse reinforcement learning. By employing this method, we learnt an individual\u27s behavioral routine preferences. We then analyzed daily routines for an individual and for eight smart home residents grouped by health diagnoses. We observed that the behavioral routine preferences changed over time. Specifically, the probability that the observed behavior was the same at the beginning of data collection as it was at the end (months later) was lower for residents experiencing cognitive decline than for cognitively healthy residents. When comparing aggregated behavior between groups of residents from the two diagnosis groups, the behavioral difference was even greater. Furthermore, the behavior preferences were used by a random forest classifier to predict a resident\u27s cognitive health diagnosis, with an accuracy of 0.84
Designing Materials Acceleration Platforms for Heterogeneous CO2 Photo(thermal)catalysis
Materials acceleration platforms (MAPs) combine automation and artificial
intelligence to accelerate the discovery of molecules and materials. They have
potential to play a role in addressing complex societal problems such as
climate change. Solar chemicals and fuels generation via heterogeneous CO2
photo(thermal)catalysis is a relatively unexplored process that holds potential
for contributing towards an environmentally and economically sustainable
future, and therefore a very promising application for MAP science and
engineering. Here, we present a brief overview of how design and innovation in
heterogeneous CO2 photo(thermal)catalysis, from materials discovery to
engineering and scale-up, could benefit from MAPs. We discuss relevant design
and performance descriptors and the level of automation of state-of-the-art
experimental techniques, and we review examples of artificial intelligence in
data analysis. Based on these precedents, we finally propose a MAP outline for
autonomous and accelerated discoveries in the emerging field of solar chemicals
and fuels sourced from CO2 photo(thermal)catalysis
Inductive queries for a drug designing robot scientist
It is increasingly clear that machine learning algorithms need to be integrated in an iterative scientific discovery loop, in which data is queried repeatedly by means of inductive queries and where the computer provides guidance to the experiments that are being performed. In this chapter, we summarise several key challenges in achieving this integration of machine learning and data mining algorithms in methods for the discovery of Quantitative Structure Activity Relationships (QSARs). We introduce the concept of a robot scientist, in which all steps of the discovery process are automated; we discuss the representation of molecular data such that knowledge discovery tools can analyse it, and we discuss the adaptation of machine learning and data mining algorithms to guide QSAR experiments
The Smart in Smart Cities: A Framework for Image Classification Using Deep Learning
The need for a smart city is more pressing today due to the recent pandemic, lockouts, climate changes, population growth, and limitations on availability/access to natural resources. However, these challenges can be better faced with the utilization of new technologies. The zoning design of smart cities can mitigate these challenges. It identifies the main components of a new smart city and then proposes a general framework for designing a smart city that tackles these elements. Then, we propose a technology-driven model to support this framework. A mapping between the proposed general framework and the proposed technology model is then introduced. To highlight the importance and usefulness of the proposed framework, we designed and implemented a smart image handling system targeted at non-technical personnel. The high cost, security, and inconvenience issues may limit the cities’ abilities to adopt such solutions. Therefore, this work also proposes to design and implement a generalized image processing model using deep learning. The proposed model accepts images from users, then performs self-tuning operations to select the best deep network, and finally produces the required insights without any human intervention. This helps in automating the decision-making process without the need for a specialized data scientist
- …