Authors: Stevanović, Sanja 
Dashti, Husain
Milošević, Marko
Al-Yakoob, Salem
Stevanović, Dragan 
Affiliations: Computer Science 
Mathematical Institute of the Serbian Academy of Sciences and Arts 
Title: Comparison of ANN and XGBoost surrogate models trained on small numbers of building energy simulations
Journal: PLoS ONE
Volume: 19
Issue: 10
First page: e0312573
Editors: Zhang, Jie
Issue Date: 1-Oct-2024
Rank: ~M22
ISSN: 1932-6203
DOI: 10.1371/journal.pone.0312573
Abstract: 
Surrogate optimisation holds a big promise for building energy optimisation studies due to its goal to replace the use of lengthy building energy simulations within an optimisation step with expendable local surrogate models that can quickly predict simulation results. To be useful for such purpose, it should be possible to quickly train precise surrogate models from a small number of simulation results (10-100) obtained from appropriately sampled points in the desired part of the design space. Two sampling methods and two machine learning models are compared here. Latin hypercube sampling (LHS), widely accepted in building energy community, is compared to an exploratory Monte Carlo-based sequential design method mc-intersite-proj-th (MIPT). Artificial neural networks (ANN), also widely accepted in building energy community, are compared to gradient-boosted tree ensembles (XGBoost), model of choice in many machine learning competitions. In order to get a better understanding of the behaviour of these two sampling methods and two machine learning models, we compare their predictions against a large set of generated synthetic data. For this purpose, a simple case study of an office cell model with a single window and a fixed overhang, whose main input parameters are overhang depth and height, while climate type, presence of obstacles, orientation and heating and cooling set points are additional input parameters, was extensively simulated with EnergyPlus, to form a large underlying dataset of 729,000 simulation results. Expendable local surrogate models for predicting simulated heating, cooling and lighting loads and equivalent primary energy needs of the office cell were trained using both LHS and MIPT and both ANN and XGBoost for several main hyperparameter choices. Results show that XGBoost models are more precise than ANN models, and that for both machine learning models, the use of MIPT sampling leads to more precise surrogates than LHS.
Publisher: Public Library of Science
Project: Science Fund of the Republic of Serbia, grant #6767, Lazy walk counts and spectral radius of threshold graphs—LZWK

Files in This Item:
File Description SizeFormat
SStevanovic.pdf3.92 MBAdobe PDFView/Open
Show full item record

Page view(s)

5
checked on Dec 22, 2024

Download(s)

3
checked on Dec 22, 2024

Google ScholarTM

Check

Altmetric

Altmetric


This item is licensed under a Creative Commons License Creative Commons