Deep Learning for Time-Series Forecasting With Exogenous Variables in Energy Consumption: A Performance and Interpretability Analysis

The rise of decentralized energy sources and renewables demands advanced grid planning, with short-term load forecasting (STLF) playing a crucial role. Energy demand in smart grids is highly variable and influenced by external factors, making accurate forecasting challenging. While deep learning mod...

Full description

Saved in:
Bibliographic Details
Main Authors: Ava Mohammadi, Gonzalo Napoles, Yamisleydi Salgueiro
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/11005580/
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The rise of decentralized energy sources and renewables demands advanced grid planning, with short-term load forecasting (STLF) playing a crucial role. Energy demand in smart grids is highly variable and influenced by external factors, making accurate forecasting challenging. While deep learning models excel in time-series forecasting, their ability to integrate exogenous variables remains uncertain. This study evaluates different deep learning architectures for STLF, including recurrent (Long Short-Term Memory, LSTM), probabilistic (Deep Autoregressive, DeepAR), attention-based (Temporal Fusion Transformer, TFT), and foundation models (TimeLLM), each designed to capture temporal dependencies differently. The “Smart Meters in London” dataset, comprising 1.6 million energy consumption records enriched with weather and socio-demographic data, was used for evaluation. Results show that incorporating exogenous variables reduces forecasting error, with TFT achieving a test MAE of 1.643, RMSE of 2.903, and sMAPE of 18.8% on a lower-variability subset, leveraging long-range dependencies for enhanced predictions. DeepAR outperformed other models on the larger, noisier dataset, achieving a test MAE of 1.699, RMSE of 3.310, and sMAPE of 18.8%, demonstrating strong generalization. LSTM performed reasonably well but struggled to utilize exogenous information, leading to higher forecasting errors. TimeLLM, despite tailored prompting for time-series forecasting, failed to outperform task-specific models, with a test MAE of 2.193, RMSE of 3.473, and sMAPE of 22.89%, highlighting the challenges of adapting foundation models trained on different modalities to structured time-series forecasting. Beyond predictive accuracy, interpretability analysis challenges the assumption that attention mechanisms reliably capture feature importance. Performance degradation analysis revealed that perturbing features ranked highly by SHAP values and permutation feature importance led to sharper error increases than attention-based rankings, with permutation demonstrating the strongest correlation with prediction errors. These findings underscore the role of Explainable AI (XAI) in time-series forecasting, emphasizing the need for robust interpretability frameworks to ensure transparency in deep learning-based predictions.
ISSN:2169-3536