News-Driven Load Forecasting: Generative Agents and Large Language Models for Unstructured Data and Event Analysis

Xinlei Wang , Jinjin Gu , Jing Qiu , Guolong Liu , Xinlei Cai , Jinzhou Zhu , Yanli Liu , Zhaoyang Dong , Junhua Zhao

Engineering ›› : 202602031

PDF
Engineering ›› :202602031 DOI: 10.1016/j.eng.2026.02.031
Article
research-article
News-Driven Load Forecasting: Generative Agents and Large Language Models for Unstructured Data and Event Analysis
Author information +
History +
PDF

Abstract

This study proposes a novel approach, intelligent text-analytic load forecasting (ITA-LF), to address the short-term load forecasting (STLF) problem by utilizing large language models (LLMs) and generative agents. This emphasizes the challenges faced by traditional forecasting methods in adapting to rapid changes and complex patterns in energy consumption, particularly during unexpected social events. It processes diverse unstructured data (e.g., historical loads, news, calendar dates, and weather), fine-tuning an LLM to enhance prediction accuracy and adaptability. An LLM-based agent with reasoning capabilities is introduced to select and understand relevant news, demonstrating the model’s ability to integrate diverse information for more precise forecasting. Our results surpass all baseline models in predictive accuracy, indicating that LLMs excel in managing the complexities of load forecasting patterns. This innovative approach not only improves forecasting accuracy but also indicates potential shifts in STLF paradigms by integrating unstructured data through advanced artificial intelligence (AI) techniques.

Keywords

Large language model / Load forecasting / Time series forecasting / Power systems

Cite this article

Download citation ▾
Xinlei Wang, Jinjin Gu, Jing Qiu, Guolong Liu, Xinlei Cai, Jinzhou Zhu, Yanli Liu, Zhaoyang Dong, Junhua Zhao. News-Driven Load Forecasting: Generative Agents and Large Language Models for Unstructured Data and Event Analysis. Engineering 202602031 DOI:10.1016/j.eng.2026.02.031

登录浏览全文

4963

注册一个新账户 忘记密码

References

[1]

Gross G, Galiana FD. Short-term load forecasting. Proc IEEE 1987; 75(12):1558-73.

[2]

Zhao WX, Zhou K, Li J, Tang T, Wang X, Hou Y, et al. A survey of large language models. 2023. arXiv:2303.18223.

[3]

Papalexopoulos AD, Hesterberg TC. A regression-based approach to short-term system load forecasting. IEEE Trans Power Syst 1990; 5(4):1535-47.

[4]

Huang SJ, Shih KR. Short-term load forecasting via ARMA model identification including non-gaussian process considerations. IEEE Trans Power Syst 2003; 18(2):673-9.

[5]

Christiaanse W. Short-term load forecasting using general exponential smoothing. IEEE Trans Power Apparatus Syst 1971; 90(2):900-11.

[6]

Chen BJ, Chang MW, Lin CJ. Load forecasting using support vector machines: a study on EUNITE competition 2001. IEEE Trans Power Syst 2004; 19(4):1821-30.

[7]

Dudek G.Short-term load forecasting using random forests. In:Proceedings of the 7th IEEE International Conference Intelligent Systems; 2014 Sep 24-26; Warsaw, Poland. Springer; 2015. p. 821-8.

[8]

Hippert HS, Pedreira CE, Souza RC. Neural networks for short-term load forecasting: a review and evaluation. IEEE Trans Power Syst 2001; 16(1):44-55.

[9]

Torres JF, Hadjout D, Sebaa A, Martínez-Álvarez F, Troncoso A. Deep learning for time series forecasting: a survey. Big Data 2021; 9(1):3-21.

[10]

Xia M, Shao H, Ma X, de Silva CW. A stacked GRU-RNN-based approach for predicting renewable energy and electricity load for smart grid operation. IEEE Trans Industr Inform 2021; 17(10):7050-9.

[11]

Liu C, Jin Z, Gu J, Qiu C.Short-term load forecasting using a long short-term memory network. In:Proceedings of 2017 IEEE PES Innovative Smart Grid Technologies Conference Europe (ISGT-Europe); 2017 Sep 26-29; IEEE; 2017. p. 1-6.

[12]

Bedi J. Attention based mechanism for load time series forecasting:AN-LSTM. In: Proceedings of Artificial Neural Networks and Machine Learning; 2020 Sep 15-18; Bratislava, Slovakia. Springer; 2020. p. 838-49.

[13]

Qu K, Si G, Shan Z, Wang Q, Liu X,Yang C. Forwardformer: efficient transformer with multi-scale forward self-attention for day-ahead load forecasting. IEEE Trans Power Syst 2023.

[14]

Cecchini M, Aytug H, Koehler GJ, Pathak P. Making words work: using financial text as a predictor of financial events. Decis Support Syst 2010; 50(1):164-75.

[15]

Schumaker RP, Chen H. Textual analysis of stock market prediction using breaking financial news: the AZFin text system. ACM Trans Inf Syst 2009; 27(2):1-19.

[16]

Schumaker RP, Chen H. A discrete stock price prediction engine based on financial news. Computer 2010; 43(1):51-6.

[17]

Bai Y, Camal S, Michiorri A. News and load: a quantitative exploration of natural language processing applications for forecasting day-ahead electricity system demand. IEEE Trans Power Syst 2024; 39(5):6222-34.

[18]

Obst D, de Vilmarest J, Goude Y. Adaptive methods for short-term electricity load forecasting during COVID-19 lockdown in France. IEEE Trans Power Syst 2021; 36(5):4754-63.

[19]

Zhou Z, Ren C, Xu Y. Forecasting masked-load with invisible distributed energy resources based on transfer learning and Bayesian tuning. Energy Convers Econ 2024; 5(5):316-26.

[20]

Joshi M, Das D, Gimpel K, Smith NA. Movie reviews and revenues:an experiment in text regression. In: Proceedings of Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics; 2010 Jun; Los Angeles, CA, USA. ACL; 2010. p. 293-6.

[21]

Zhang X, Saleh H, Younis EM, Sahal R, Ali AA. Predicting coronavirus pandemic in real-time using machine learning and big data streaming system. Complexity 2020; 2020:1-10.

[22]

Rodrigues F, Markou I, Pereira FC. Combining time-series and textual data for taxi demand prediction in event areas: a deep learning approach. Inf Fusion 2019; 49:120-9.

[23]

Chowdhary K.Natural language processing. In:Fundamentals of artificial intelligence. New Delhi:Springer; 2020. p. 603-49.

[24]

Berrios W, Mittal G, Thrush T, Kiela D,Singh A. Towards language models that can see: computer vision through the lens of natural language. 2023. arXiv:2306.16410.

[25]

Ghosal D, Majumder N, Mehrish A,Poria S. Text-to-audio generation using instruction-tuned LLM and latent diffusion model. 2023. arXiv:2304.13731.

[26]

Jin M, Wen Q, Liang Y, Zhang C, Xue S, Wang X, et al. Large models for time series and spatio-temporal data: a survey and outlook. 2023. arXiv:2310.10196.

[27]

Jin M, Wang S, Ma L, Chu Z, Zhang JY, Shi X, et al. Time-LLM: time series forecasting by reprogramming large language models. 2023. arXiv:2310.01728.

[28]

Garza A,Mergenthaler-Canseco M. TimeGPT-1. 2023. arXiv:2310.03589.

[29]

Yu X, Chen Z, Ling Y, Dong S, Liu Z,Lu Y. Temporal data meets LLM-explainable financial time series forecasting. 2023. arXiv:2306.11025.

[30]

Xue H, Voutharoja BP, Salim FD. 2022 Nov 1-4; Seattle WA, Leveraging language foundation models for human mobility forecasting. Proceedings of the 30th international conference on advances in geographic information systems; USA. ACM; 2022. p. 1-9.

[31]

Wang X, Fang M, Zeng Z, Cheng T.Where would I go next? Large language models as human mobility predictors. 2023. arXiv:2308.15197.

[32]

Liu X, McDuff D, Kovacs G, Galatzer-Levy I, Sunshine J, Zhan J, et al. Large language models are few-shot health learners. 2023. arXiv:2305.15525.

[33]

Karim SA, Alwi SA. Electricity load forecasting in UTP using moving averages and exponential smoothing techniques. Appl Math Sci 2013; 7(77-80):4003-14.

[34]

Ryu S, Noh J, Kim H. Deep neural network based demand side short term load forecasting. Energies 2016; 10(1):3.

[35]

Sadaei HJ, de Lima e Silva PC, Guimarães FG, Lee MH. Short-term load forecasting by using a combined method of convolutional neural networks and fuzzy time series. Energy 2019; 175:365-77.

[36]

Wu H, Hu T, Liu Y, Zhou H, Wang J, Long M.TimesNet:Temporal 2D-variation modeling for general time series analysis. In: Proceedings of Eleventh International Conference on Learning Representations; 2022 May 1-5; Kigali, Rwanda. ICLR; 2022.

[37]

Kong W, Dong ZY, Jia Y, Hill DJ, Xu Y, Zhang Y. Short-term residential load forecasting based on LSTM recurrent neural network. IEEE Trans Smart Grid 2017; 10(1):841-51.

[38]

Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. 2017. arXiv:1706.03762.

[39]

Du J, Cheng Y, Zhou Q, Zhang J, Zhang X, Li G. Power load forecasting using BiLSTM-attention. IOP Conf Ser Earth Environ Sci 2020; 440:032115.

[40]

Zhou H, Zhang S, Peng J, Zhang S, Li J, Xiong H, et al. Informer: beyond efficient transformer for long sequence time-series forecasting. AAAI Conf Artif Intell 2021; 35(12):11106-15.

[41]

Wu H, Xu J, Wang J, Long M. Autoformer: decomposition transformers with auto-correlation for long-term series forecasting. Adv Neural Inf Process Syst 2021; 34:22419-30.

[42]

Zhou T, Ma Z, Wen Q, Wang X, Sun L, Jin R. FEDformer:frequency enhanced decomposed transformer for long-term series forecasting. In: Proceedings of International conference on machine learning; 2022 Jul 17-23; Baltimore, MD, USA. PMLR; 2022. p. 27268-86.

[43]

Liu S, Yu H, Liao C, Li J, Lin W, Liu AX, et al. Pyraformer:low-complexity pyramidal attention for long-range time series modeling and forecasting. In: Proceedings of International conference on learning representations; 2022 Apr 25; online. OpenReview; 2022.

[44]

Nie Y, Nguyen NH, Sinthong P, Kalagnanam J. A time series is worth 64 words: long-term forecasting with transformers. 2022. arXiv:2211.14730.

[45]

Liu Y, He T, Zhang H, Wu H, Wang S, Ma L, et al. iTransformer: inverted transformers are effective for time series forecasting. 2023. arXiv:2310.06625.

[46]

Zhou T, Ma Z, Wang X, Wen Q, Sun L, Yao T, et al. FiLM: frequency improved Legendre memory model for long-term time series forecasting. Adv Neural Inf Process Syst 2022; 35:12677-12690.

[47]

Zhao Z, Xia C, Chi L, Chang X, Li W, Yang T, et al. Short-term load forecasting based on the transformer model. Information 2021; 12(12):516.

[48]

Wang C, Wang Y, Ding Z, Zheng T, Hu J, Zhang K. A transformer-based method of multienergy load forecasting in integrated energy system. IEEE Trans Smart Grid 2022; 13(4):2703-14.

[49]

Ryu S, Yu Y. Quantile-mixer: a novel deep learning approach for probabilistic short-term load forecasting. IEEE Trans Smart Grid 2024; 15(2):2237-50.

[50]

Taylor SJ, Letham B. Forecasting at scale. Am Stat 2018; 72(1):37-45.

[51]

Devlin J, Chang MW, Lee K,Toutanova K. Bert: pre-training of deep bidirectional transformers for language understanding. 2018. arXiv:1810.04805.

[52]

Lewis M, Liu Y, Goyal N, Ghazvininejad M, Mohamed A, Levy O, et al. BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. 2019. arXiv:1910.13461.

[53]

GPT-4Vision system card. Report. San Francisco: OpenAI; 2023.

[54]

Xue H,Salim FD. Prompt-based time series forecasting: a new task and dataset. 2022. arXiv:2210.08964.

[55]

Wei J, Wang X, Schuurmans D, Bosma M, Ichter B, Xia F, et al. Chain-of-thought prompting elicits reasoning in large language models. Adv Neural Inf Proc Syst 2022; 35:24824-37.

[56]

Shinn N, Labash B,Gopinath A. Reflexion: an autonomous agent with dynamic memory and self-reflection. 2023. arXiv:2303.11366.

[57]

Cai T, Wang X, Ma T, Chen X,Zhou D. Large language models as tool makers. 2023. arXiv:2305.17126.

[58]

Yao S, Zhao J, Yu D, Du N, Shafran I, Narasimhan K, et al. React: synergizing reasoning and acting in language models. 2022. arXiv:2210.03629.

[59]

Park JS, O’Brien J, Cai CJ, Morris MR, Liang P, Bernstein MS. Generative agents: Interactive simulacra of human behavior. In: Proceedings of the 36th annual ACM symposium on user interface software and technology; 2023 Oct 29-Nov 1; San Francisco, CA, USA. ACM; 2023. p. 1-22.

[60]

Du Y, Li S, Torralba A, Tenenbaum JB,Mordatch I. Improving factuality and reasoning in language models through multiagent debate. 2023. arXiv:2305.14325.

[61]

Touvron H, Lavril T, Izacard G, Martinet X, Lachaux MA, Lacroix T, et al. LLaMA: open and efficient foundation language models. 2023. arXiv:2302.13971.

[62]

Elfwing S, Uchibe E, Doya K. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw 2018; 107:3-11.

[63]

Zhang B, Sennrich R.Root mean square layer normalization. In: Proceedings of Advances in Neural Information Processing Systems; 2019 Dec 8-14; Vancouver, BC, Canada. University Library of Zurich; 2019.

[64]

Su J, Ahmed M, Lu Y, Pan S, Bo W, Liu Y. Roformer: enhanced transformer with rotary position embedding. Neurocomputing 2024; 568:127063.

[65]

Hu EJ, Shen Y, Wallis P, Allen-Zhu Z, Li Y, Wang S, et al. LoRA: low-rank adaptation of large language models. 2021. arXiv:2106.09685.

[66]

Touvron H, Martin L, Stone K, Albert P, Almahairi A, Babaei Y, et al. Llama 2: open foundation and fine-tuned chat models. 2023. arXiv:2307.09288.

[67]

Zeng A, Chen M, Zhang L, Xu Q.Are transformers effective for time series forecasting? In: Proceedings of the AAAI Conference on Artificial Intelligence; 2023 Feb 7-14; Washington, DC, USA. Association for the Advancement of Artificial Intelligence; 2023. p. 11121-8.

[68]

Zhou T, Niu P, Wang X, Sun L, Jin R. One fits all: power general time series analysis by pretrained LM. Adv Neural Inf Process Syst 2023; 36:43322-55.

PDF

0

Accesses

0

Citation

Detail

Sections
Recommended

/