Prediction of aerosol extinction of coefficient in coastal areas of South China based on informer-LSTF

Zhou Ye, Shengcheng Cui, Zhi-li Qiao, Huiqiang Xu, Zihan Zhang, Tao Luo, Xuebin Li
{"title":"Prediction of aerosol extinction of coefficient in coastal areas of South China based on informer-LSTF","authors":"Zhou Ye, Shengcheng Cui, Zhi-li Qiao, Huiqiang Xu, Zihan Zhang, Tao Luo, Xuebin Li","doi":"10.1117/12.2681810","DOIUrl":null,"url":null,"abstract":"Models related to long and short-term memory networks have demonstrated superior performance in short-term prediction, but their prediction ability becomes limited in long sequence time series forecasting (LSTF), and prediction time increases. To address these issues, this paper optimizes the Transformer and Informer models in the following ways: (1) input representation optimization, by adding a time embedding layer representing global timestamps and a positional embedding layer to improve the model's prediction ability for aerosol extinction coefficient (AEC); (2) self-attention mechanism optimization, by using probabilistic self-attention mechanism and self-attention distillation mechanism to reduce memory usage and enhance the model's local modeling ability through convolutional aggregation operations; (3) generative decoding, using dynamic decoding to enhance the model's long sequence prediction ability. Based on these optimizations, a new LSTF model for AEC is proposed in this paper. Experimental results on the atmosphere parameters of the Maoming (APM) dataset and weather dataset show that the proposed model has significant improvements in accuracy, memory usage, and runtime speed compared to other similar Transformer models. In the accuracy experiment, compared to the Transformer model, the MAE of this model on APM dataset decreased from 0.237 to 0.103, and the MSE decreased from 0.345 to 241. In the memory usage experiment, the model can effectively alleviate memory overflow problems when the input length is greater than 720. In the runtime speed experiment, when the input length is 672, the training time per round decreased from 15.32 seconds to 12.39 seconds. These experiments demonstrate the effectiveness and reliability of the proposed model, providing a new approach and method for long sequence prediction of AEC.","PeriodicalId":130374,"journal":{"name":"Semantic Ambient Media Experiences","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Semantic Ambient Media Experiences","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.2681810","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Models related to long and short-term memory networks have demonstrated superior performance in short-term prediction, but their prediction ability becomes limited in long sequence time series forecasting (LSTF), and prediction time increases. To address these issues, this paper optimizes the Transformer and Informer models in the following ways: (1) input representation optimization, by adding a time embedding layer representing global timestamps and a positional embedding layer to improve the model's prediction ability for aerosol extinction coefficient (AEC); (2) self-attention mechanism optimization, by using probabilistic self-attention mechanism and self-attention distillation mechanism to reduce memory usage and enhance the model's local modeling ability through convolutional aggregation operations; (3) generative decoding, using dynamic decoding to enhance the model's long sequence prediction ability. Based on these optimizations, a new LSTF model for AEC is proposed in this paper. Experimental results on the atmosphere parameters of the Maoming (APM) dataset and weather dataset show that the proposed model has significant improvements in accuracy, memory usage, and runtime speed compared to other similar Transformer models. In the accuracy experiment, compared to the Transformer model, the MAE of this model on APM dataset decreased from 0.237 to 0.103, and the MSE decreased from 0.345 to 241. In the memory usage experiment, the model can effectively alleviate memory overflow problems when the input length is greater than 720. In the runtime speed experiment, when the input length is 672, the training time per round decreased from 15.32 seconds to 12.39 seconds. These experiments demonstrate the effectiveness and reliability of the proposed model, providing a new approach and method for long sequence prediction of AEC.
基于inform - lstf的华南沿海气溶胶消光系数预测
长短期记忆网络模型在短期预测中表现出较好的性能,但在长序列时间序列预测中,其预测能力受到限制,预测时间增加。针对这些问题,本文对Transformer和Informer模型进行了如下优化:(1)输入表示优化,通过增加表示全局时间戳的时间嵌入层和位置嵌入层,提高模型对气溶胶消光系数(AEC)的预测能力;(2)自注意机制优化,利用概率自注意机制和自注意蒸馏机制,通过卷积聚合操作减少内存使用,增强模型的局部建模能力;(3)生成式解码,采用动态解码增强模型的长序列预测能力。在此基础上,提出了一种新的AEC LSTF模型。在茂名(APM)数据集和天气数据集的大气参数实验结果表明,与其他类似的Transformer模型相比,该模型在精度、内存占用和运行速度方面都有显著提高。在精度实验中,与Transformer模型相比,该模型在APM数据集上的MAE从0.237降低到0.103,MSE从0.345降低到241。在内存使用实验中,该模型可以有效缓解输入长度大于720时的内存溢出问题。在运行速度实验中,当输入长度为672时,每轮训练时间从15.32秒减少到12.39秒。实验验证了该模型的有效性和可靠性,为AEC的长序列预测提供了一种新的途径和方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信