Transfer learning on transformers for building energy consumption forecasting—A comparative study

Loading...
Thumbnail Image

Date

2025-06-01

DOI

Open Access Location

Journal Title

Journal ISSN

Volume Title

Publisher

Elsevier B V

Rights

(c) The author/s
CC BY

Abstract

Energy consumption in buildings is steadily increasing, leading to higher carbon emissions. Predicting energy consumption is a key factor in addressing climate change. There has been a significant shift from traditional statistical models to advanced deep learning (DL) techniques for predicting energy use in buildings. However, data scarcity in newly constructed or poorly instrumented buildings limits the effectiveness of standard DL approaches. In this study, we investigate the application of six data-centric Transfer Learning (TL) strategies on three Transformer architectures—vanilla Transformer, Informer, and PatchTST—to enhance building energy consumption forecasting. Transformers, a relatively new DL framework, have demonstrated significant promise in various domains; yet, prior TL research has often focused on either a single data-centric strategy or older models such as Recurrent Neural Networks. Using 16 diverse datasets from the Building Data Genome Project 2, we conduct an extensive empirical analysis under varying feature spaces (e.g., recorded ambient weather) and building characteristics (e.g., dataset volume). Our experiments show that combining multiple source datasets under a zero-shot setup reduces the Mean Absolute Error (MAE) of the vanilla Transformer model by an average of 15.9 % for 24 h forecasts, compared to single-source baselines. Further fine-tuning these multi-source models with target-domain data yields an additional 3–5 % improvement. Notably, PatchTST outperforms the vanilla Transformer and Informer models. Overall, our results underscore the potential of combining Transformer architectures with TL techniques to enhance building energy consumption forecasting accuracy. However, careful selection of the TL strategy and attention to feature space compatibility are needed to maximize forecasting gains.

Description

Keywords

Building energy consumption forecasting, Transfer learning for time series, Transformer models for time series forecasting, Data-centric transfer learning strategies, PatchTST, Informer, Zero-shot learning, Model fine-tuning, Data scarcity

Citation

Spencer R, Ranathunga S, Boulic M, van Heerden AH, Susnjak T. (2025). Transfer learning on transformers for building energy consumption forecasting—A comparative study. Energy and Buildings. 336.

Collections

Endorsement

Review

Supplemented By

Referenced By

Creative Commons license

Except where otherwised noted, this item's license is described as (c) The author/s