Deep learning Calabi-Yau four folds with hybrid and recurrent neural network architectures

IF 2.5 3区 物理与天体物理 Q2 PHYSICS, PARTICLES & FIELDS
Harriet L. Dao
{"title":"Deep learning Calabi-Yau four folds with hybrid and recurrent neural network architectures","authors":"Harriet L. Dao","doi":"10.1016/j.nuclphysb.2025.116832","DOIUrl":null,"url":null,"abstract":"<div><div>In this work, we report the results of applying deep learning based on hybrid convolutional-recurrent and purely recurrent neural network architectures to the dataset of almost one million complete intersection Calabi-Yau four-folds (CICY4) to machine-learn their four Hodge numbers <span><math><msup><mrow><mi>h</mi></mrow><mrow><mn>1</mn><mo>,</mo><mn>1</mn></mrow></msup></math></span>, <span><math><msup><mrow><mi>h</mi></mrow><mrow><mn>2</mn><mo>,</mo><mn>1</mn></mrow></msup></math></span>, <span><math><msup><mrow><mi>h</mi></mrow><mrow><mn>3</mn><mo>,</mo><mn>1</mn></mrow></msup></math></span>, <span><math><msup><mrow><mi>h</mi></mrow><mrow><mn>2</mn><mo>,</mo><mn>2</mn></mrow></msup></math></span>. In particular, we explored and experimented with twelve different neural network models, nine of which are convolutional-recurrent (CNN-RNN) hybrids with the RNN unit being either GRU (Gated Recurrent Unit) or Long Short Term Memory (LSTM). The remaining four models are purely recurrent neural networks based on LSTM. In terms of the <span><math><msup><mrow><mi>h</mi></mrow><mrow><mn>1</mn><mo>,</mo><mn>1</mn></mrow></msup></math></span>, <span><math><msup><mrow><mi>h</mi></mrow><mrow><mn>2</mn><mo>,</mo><mn>1</mn></mrow></msup></math></span>, <span><math><msup><mrow><mi>h</mi></mrow><mrow><mn>3</mn><mo>,</mo><mn>1</mn></mrow></msup></math></span> and <span><math><msup><mrow><mi>h</mi></mrow><mrow><mn>2</mn><mo>,</mo><mn>2</mn></mrow></msup></math></span> prediction accuracies, at 72% training ratio, our best performing individual model is CNN-LSTM-400, a hybrid CNN-LSTM with the LSTM hidden size of 400, which obtained 99.74%, 98.07%, 95.19%, 81.01%, our second best performing individual model is LSTM-448, an LSTM-based model with the hidden size of 448, which obtained 99.74%, 97.51%, 94.24%, and 78.63%. These results were improved by forming ensembles of the top two, three or even four models. Our best ensemble, consisting of the top four models, achieved the accuracies of 99.84%, 98.71%, 96.26%, 85.03%. At 80% training ratio, the top two performing models LSTM-448 and LSTM-424 are both LSTM-based with the hidden sizes of 448 and 424. Compared with the 72% training ratio, there is a significant improvement of accuracies, which reached 99.85%, 98.66%, 96.26%, 84.77% for the best individual model and 99.90%, 99.03%, 97.97%, 87.34% for the best ensemble. By nature a proof of concept, the results of this work conclusively established the utility of RNN-based architectures and demonstrated their effective performances compared to the well-explored purely CNN-based architectures in the problem of deep learning Calabi Yau manifolds.</div></div>","PeriodicalId":54712,"journal":{"name":"Nuclear Physics B","volume":"1013 ","pages":"Article 116832"},"PeriodicalIF":2.5000,"publicationDate":"2025-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Nuclear Physics B","FirstCategoryId":"101","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0550321325000422","RegionNum":3,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"PHYSICS, PARTICLES & FIELDS","Score":null,"Total":0}
引用次数: 0

Abstract

In this work, we report the results of applying deep learning based on hybrid convolutional-recurrent and purely recurrent neural network architectures to the dataset of almost one million complete intersection Calabi-Yau four-folds (CICY4) to machine-learn their four Hodge numbers h1,1, h2,1, h3,1, h2,2. In particular, we explored and experimented with twelve different neural network models, nine of which are convolutional-recurrent (CNN-RNN) hybrids with the RNN unit being either GRU (Gated Recurrent Unit) or Long Short Term Memory (LSTM). The remaining four models are purely recurrent neural networks based on LSTM. In terms of the h1,1, h2,1, h3,1 and h2,2 prediction accuracies, at 72% training ratio, our best performing individual model is CNN-LSTM-400, a hybrid CNN-LSTM with the LSTM hidden size of 400, which obtained 99.74%, 98.07%, 95.19%, 81.01%, our second best performing individual model is LSTM-448, an LSTM-based model with the hidden size of 448, which obtained 99.74%, 97.51%, 94.24%, and 78.63%. These results were improved by forming ensembles of the top two, three or even four models. Our best ensemble, consisting of the top four models, achieved the accuracies of 99.84%, 98.71%, 96.26%, 85.03%. At 80% training ratio, the top two performing models LSTM-448 and LSTM-424 are both LSTM-based with the hidden sizes of 448 and 424. Compared with the 72% training ratio, there is a significant improvement of accuracies, which reached 99.85%, 98.66%, 96.26%, 84.77% for the best individual model and 99.90%, 99.03%, 97.97%, 87.34% for the best ensemble. By nature a proof of concept, the results of this work conclusively established the utility of RNN-based architectures and demonstrated their effective performances compared to the well-explored purely CNN-based architectures in the problem of deep learning Calabi Yau manifolds.
求助全文
约1分钟内获得全文 求助全文
来源期刊
Nuclear Physics B
Nuclear Physics B 物理-物理:粒子与场物理
CiteScore
5.50
自引率
7.10%
发文量
302
审稿时长
1 months
期刊介绍: Nuclear Physics B focuses on the domain of high energy physics, quantum field theory, statistical systems, and mathematical physics, and includes four main sections: high energy physics - phenomenology, high energy physics - theory, high energy physics - experiment, and quantum field theory, statistical systems, and mathematical physics. The emphasis is on original research papers (Frontiers Articles or Full Length Articles), but Review Articles are also welcome.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信