[Poster Presentation]GPT-2-Augmented Sequence Modeling for Short-Term Load Forecasting

GPT-2-Augmented Sequence Modeling for Short-Term Load Forecasting
ID:77 Submission ID:82 View Protection:ATTENDEE Updated Time:2025-10-11 22:47:12 Hits:48 Poster Presentation

Start Time:2025-11-09 09:07 (Asia/Shanghai)

Duration:1min

Session:[P] Poster presentation » [P6] 6.AI-driven technology

No files

Abstract
Abstract -- Load forecasting serves as the foundation for power system operation and planning. Accurate load forecasting ensures the secure and reliable operation of power systems, reduces generation costs, and enhances economic efficiency. Recent studies demonstrate that large language models (LLMs) exhibit powerful capabilities in pattern recognition and reasoning for complex token sequences. The critical challenge lies in effectively aligning temporal patterns in time-series data with linguistic structures in natural language to leverage these capabilities. This paper proposes a large model-based time-series forecasting method for electrical load prediction. The approach leverages a pre-trained GPT-2 (Generative Pre-trained Transformer 2) model as its foundation while freezing parameters in its self-attention and feed-forward neural network layers. Fine-tuning is applied exclusively to the input embedding layer and output projection layer. Experimental results demonstrate that the proposed method achieves performance comparable to or superior against existing approaches across multiple electrical load forecasting tasks.
 
Keywords
Load forecasting, time-series data, LLM, GPT-2, Fine-tuning
Speaker
Xu Kun
graduate student Southeast university

Submission Author
Xu Kun Southeast university
Wang Ying Southeast University
Comment submit
Verification code Change another
All comments

Contact us

CIYCEE 2025 Official E-mail:

ciycee2025@163.com

 

WeChat public account: 

IEEE IAS SWJTU Student Branch