{"title":"基于扩展多尺度变压器的鲁棒高效光伏电力预测","authors":"Yingying Liu, Zhongping Wang","doi":"10.1016/j.solener.2025.114077","DOIUrl":null,"url":null,"abstract":"<div><div>Accurate and reliable multi-horizon photovoltaic (PV) power forecasting is essential for seamless grid integration, economic dispatch, and power system stability. However, the inherent variability of solar irradiance and complex nonlinear dynamics pose significant challenges. To address these issues, this study proposes WinFlashDilated-former, a model specifically designed for efficient LSTF in the PV domain.Built upon an encoder-decoder architecture, the model introduces three key innovations to overcome the limitations of existing methods:(1) a precise multi-scale window self-attention mechanism deeply optimized with FlashAttention-3, which captures multi-scale dependencies, from short-term fluctuations to long-term trends, without loss of information while achieving excellent computational efficiency;(2) a dilated convolutional feed-forward network in the encoder, which expands the receptive field to enhance modeling of long-range contextual information; and (3) a time-enhanced windowed autoregressive generation strategy in the decoder, which strengthens the perception of periodic patterns by leveraging future temporal attributes.Extensive experiments were conducted on real-world datasets provided by DKASC, covering multiple PV technologies and multi-year operational records. The results demonstrate that WinFlashDilated-former consistently outperforms both traditional time-series models and several SOTA LSTF approaches across all forecasting horizons from 5-min to 12-h. In the 5-min horizon, the model achieved MAE(0.0009), RMSE(0.0146), and R<sup>2</sup>(0.9972). In the 12-h horizon, compared to the baseline method, MAE and RMSE were reduced by 13.2% and 12.03%, respectively, while maintaining strong robustness.Ablation studies further confirm the critical contributions of the multi-scale attention mechanism and dilated convolutional feed-forward network. Moreover, WinFlashDilated-former achieves an excellent balance between accuracy and efficiency, demonstrating competitive inference speed and GPU memory utilization. The code and data are publicly available at: <span><span>https://github.com/YingyingLiu2002/winflashdilated-former</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":428,"journal":{"name":"Solar Energy","volume":"302 ","pages":"Article 114077"},"PeriodicalIF":6.0000,"publicationDate":"2025-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Robust and efficient multihorizon photovoltaic power forecasting with a dilated multi-scale Transformer\",\"authors\":\"Yingying Liu, Zhongping Wang\",\"doi\":\"10.1016/j.solener.2025.114077\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Accurate and reliable multi-horizon photovoltaic (PV) power forecasting is essential for seamless grid integration, economic dispatch, and power system stability. However, the inherent variability of solar irradiance and complex nonlinear dynamics pose significant challenges. To address these issues, this study proposes WinFlashDilated-former, a model specifically designed for efficient LSTF in the PV domain.Built upon an encoder-decoder architecture, the model introduces three key innovations to overcome the limitations of existing methods:(1) a precise multi-scale window self-attention mechanism deeply optimized with FlashAttention-3, which captures multi-scale dependencies, from short-term fluctuations to long-term trends, without loss of information while achieving excellent computational efficiency;(2) a dilated convolutional feed-forward network in the encoder, which expands the receptive field to enhance modeling of long-range contextual information; and (3) a time-enhanced windowed autoregressive generation strategy in the decoder, which strengthens the perception of periodic patterns by leveraging future temporal attributes.Extensive experiments were conducted on real-world datasets provided by DKASC, covering multiple PV technologies and multi-year operational records. The results demonstrate that WinFlashDilated-former consistently outperforms both traditional time-series models and several SOTA LSTF approaches across all forecasting horizons from 5-min to 12-h. In the 5-min horizon, the model achieved MAE(0.0009), RMSE(0.0146), and R<sup>2</sup>(0.9972). In the 12-h horizon, compared to the baseline method, MAE and RMSE were reduced by 13.2% and 12.03%, respectively, while maintaining strong robustness.Ablation studies further confirm the critical contributions of the multi-scale attention mechanism and dilated convolutional feed-forward network. Moreover, WinFlashDilated-former achieves an excellent balance between accuracy and efficiency, demonstrating competitive inference speed and GPU memory utilization. The code and data are publicly available at: <span><span>https://github.com/YingyingLiu2002/winflashdilated-former</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":428,\"journal\":{\"name\":\"Solar Energy\",\"volume\":\"302 \",\"pages\":\"Article 114077\"},\"PeriodicalIF\":6.0000,\"publicationDate\":\"2025-10-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Solar Energy\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0038092X25008400\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENERGY & FUELS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Solar Energy","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0038092X25008400","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENERGY & FUELS","Score":null,"Total":0}
Robust and efficient multihorizon photovoltaic power forecasting with a dilated multi-scale Transformer
Accurate and reliable multi-horizon photovoltaic (PV) power forecasting is essential for seamless grid integration, economic dispatch, and power system stability. However, the inherent variability of solar irradiance and complex nonlinear dynamics pose significant challenges. To address these issues, this study proposes WinFlashDilated-former, a model specifically designed for efficient LSTF in the PV domain.Built upon an encoder-decoder architecture, the model introduces three key innovations to overcome the limitations of existing methods:(1) a precise multi-scale window self-attention mechanism deeply optimized with FlashAttention-3, which captures multi-scale dependencies, from short-term fluctuations to long-term trends, without loss of information while achieving excellent computational efficiency;(2) a dilated convolutional feed-forward network in the encoder, which expands the receptive field to enhance modeling of long-range contextual information; and (3) a time-enhanced windowed autoregressive generation strategy in the decoder, which strengthens the perception of periodic patterns by leveraging future temporal attributes.Extensive experiments were conducted on real-world datasets provided by DKASC, covering multiple PV technologies and multi-year operational records. The results demonstrate that WinFlashDilated-former consistently outperforms both traditional time-series models and several SOTA LSTF approaches across all forecasting horizons from 5-min to 12-h. In the 5-min horizon, the model achieved MAE(0.0009), RMSE(0.0146), and R2(0.9972). In the 12-h horizon, compared to the baseline method, MAE and RMSE were reduced by 13.2% and 12.03%, respectively, while maintaining strong robustness.Ablation studies further confirm the critical contributions of the multi-scale attention mechanism and dilated convolutional feed-forward network. Moreover, WinFlashDilated-former achieves an excellent balance between accuracy and efficiency, demonstrating competitive inference speed and GPU memory utilization. The code and data are publicly available at: https://github.com/YingyingLiu2002/winflashdilated-former.
期刊介绍:
Solar Energy welcomes manuscripts presenting information not previously published in journals on any aspect of solar energy research, development, application, measurement or policy. The term "solar energy" in this context includes the indirect uses such as wind energy and biomass