Transfer learning on transformers for building energy consumption forecasting—A comparative study
Loading...

Date
2025-06-01
Open Access Location
Journal Title
Journal ISSN
Volume Title
Publisher
Elsevier B V
Rights
(c) The author/s
CC BY
CC BY
Abstract
Energy consumption in buildings is steadily increasing, leading to higher carbon emissions. Predicting energy consumption is a key factor in addressing climate change. There has been a significant shift from traditional statistical models to advanced deep learning (DL) techniques for predicting energy use in buildings. However, data scarcity in newly constructed or poorly instrumented buildings limits the effectiveness of standard DL approaches. In this study, we investigate the application of six data-centric Transfer Learning (TL) strategies on three Transformer architectures—vanilla Transformer, Informer, and PatchTST—to enhance building energy consumption forecasting. Transformers, a relatively new DL framework, have demonstrated significant promise in various domains; yet, prior TL research has often focused on either a single data-centric strategy or older models such as Recurrent Neural Networks. Using 16 diverse datasets from the Building Data Genome Project 2, we conduct an extensive empirical analysis under varying feature spaces (e.g., recorded ambient weather) and building characteristics (e.g., dataset volume). Our experiments show that combining multiple source datasets under a zero-shot setup reduces the Mean Absolute Error (MAE) of the vanilla Transformer model by an average of 15.9 % for 24 h forecasts, compared to single-source baselines. Further fine-tuning these multi-source models with target-domain data yields an additional 3–5 % improvement. Notably, PatchTST outperforms the vanilla Transformer and Informer models. Overall, our results underscore the potential of combining Transformer architectures with TL techniques to enhance building energy consumption forecasting accuracy. However, careful selection of the TL strategy and attention to feature space compatibility are needed to maximize forecasting gains.
Description
Keywords
Building energy consumption forecasting, Transfer learning for time series, Transformer models for time series forecasting, Data-centric transfer learning strategies, PatchTST, Informer, Zero-shot learning, Model fine-tuning, Data scarcity
Citation
Spencer R, Ranathunga S, Boulic M, van Heerden AH, Susnjak T. (2025). Transfer learning on transformers for building energy consumption forecasting—A comparative study. Energy and Buildings. 336.