Any ZeaD Formula of Six Instants Having No Quartic or Higher Precision with Proof

Yunong Zhang, Jinjin Guo, Liu He, Yang Shi, Chaowei Hu
{"title":"Any ZeaD Formula of Six Instants Having No Quartic or Higher Precision with Proof","authors":"Yunong Zhang, Jinjin Guo, Liu He, Yang Shi, Chaowei Hu","doi":"10.1109/ICSAI.2018.8599451","DOIUrl":null,"url":null,"abstract":"In recent years, Zhang et al. discretization (ZeaD) as a new class of time-discretization methods has been proposed, named and applied by Zhang et al. Note that ZeaD formulas can accurately discretize Zhang neural networks $(\\mathrm {i}.\\mathrm {e}.$, ZNN, or say, Zhang dynamics) models as well as ordinary differential equation systems. In previous work, various ZeaD formulas have been presented and unified, including Euler forward formula as 2-instant ZeaD formula that is convergent with a truncation error being proportional to the first power of sampling period and Taylor-type discretization formula as 4-instant ZeaD formula that is convergent with a truncation error being proportional to the second power of sampling period. During our pursuit of ZeaD formulas that are convergent with a higher precision, we discover that there exists no 6-instant ZeaD formula that is convergent with a quartic (ie, biquadratic, of degree 4) or higher precision. The truncation error of any 6-instant ZeaD formula is proportional to the third power of sampling period or bigger. The contributions are theoretically proved in this paper as well.","PeriodicalId":375852,"journal":{"name":"2018 5th International Conference on Systems and Informatics (ICSAI)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 5th International Conference on Systems and Informatics (ICSAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSAI.2018.8599451","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

In recent years, Zhang et al. discretization (ZeaD) as a new class of time-discretization methods has been proposed, named and applied by Zhang et al. Note that ZeaD formulas can accurately discretize Zhang neural networks $(\mathrm {i}.\mathrm {e}.$, ZNN, or say, Zhang dynamics) models as well as ordinary differential equation systems. In previous work, various ZeaD formulas have been presented and unified, including Euler forward formula as 2-instant ZeaD formula that is convergent with a truncation error being proportional to the first power of sampling period and Taylor-type discretization formula as 4-instant ZeaD formula that is convergent with a truncation error being proportional to the second power of sampling period. During our pursuit of ZeaD formulas that are convergent with a higher precision, we discover that there exists no 6-instant ZeaD formula that is convergent with a quartic (ie, biquadratic, of degree 4) or higher precision. The truncation error of any 6-instant ZeaD formula is proportional to the third power of sampling period or bigger. The contributions are theoretically proved in this paper as well.
任何没有四次精度或更高精度的六阶微分公式并证明
近年来,Zhang等人提出了一种新的时间离散化方法——离散化(ZeaD),并对其进行了命名和应用。注意,ZeaD公式可以精确地离散张神经网络$(\ mathm {i})。\ mathrm {e}。$, ZNN,或者说是张动力学)模型以及常微分方程系统。在之前的工作中,已经提出并统一了各种ZeaD公式,包括欧拉正演公式为2瞬时ZeaD公式,其收敛性与截断误差与采样周期的一次方成正比,泰勒型离散化公式为4瞬时ZeaD公式,其收敛性与截断误差与采样周期的二次方成正比。在我们追求收敛精度更高的ZeaD公式的过程中,我们发现不存在收敛精度更高的四次(即双二次)的6瞬时ZeaD公式。任何6瞬时ZeaD公式的截断误差与采样周期的三次幂或更大成正比。本文也从理论上证明了这些贡献。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信