质子-质子碰撞中重强子衰变的深度全事件解释和分层重建的GNN。

Q1 Computer Science
Computing and Software for Big Science Pub Date : 2023-01-01 Epub Date: 2023-11-17 DOI:10.1007/s41781-023-00107-8
Julián García Pardiñas, Marta Calvi, Jonas Eschle, Andrea Mauri, Simone Meloni, Martina Mozzanica, Nicola Serra
{"title":"质子-质子碰撞中重强子衰变的深度全事件解释和分层重建的GNN。","authors":"Julián García Pardiñas, Marta Calvi, Jonas Eschle, Andrea Mauri, Simone Meloni, Martina Mozzanica, Nicola Serra","doi":"10.1007/s41781-023-00107-8","DOIUrl":null,"url":null,"abstract":"<p><p>The LHCb experiment at the Large Hadron Collider (LHC) is designed to perform high-precision measurements of heavy-hadron decays, which requires the collection of large data samples and a good understanding and suppression of multiple background sources. Both factors are challenged by a fivefold increase in the average number of proton-proton collisions per bunch crossing, corresponding to a change in the detector operation conditions for the LHCb Upgrade I phase, recently started. A further tenfold increase is expected in the Upgrade II phase, planned for the next decade. The limits in the storage capacity of the trigger will bring an inverse relationship between the number of particles selected to be stored per event and the number of events that can be recorded. In addition the background levels will rise due to the enlarged combinatorics. To tackle both challenges, we propose a novel approach, never attempted before in a hadronic collider: a Deep-learning based Full Event Interpretation (DFEI), to perform the simultaneous identification, isolation and hierarchical reconstruction of all the heavy-hadron decay chains per event. This strategy radically contrasts with the standard selection procedure used in LHCb to identify heavy-hadron decays, that looks individually at subsets of particles compatible with being products of specific decay types, disregarding the contextual information from the rest of the event. Following the DFEI approach, once the relevant particles in each event are identified, the rest can be safely removed to optimise the storage space and maximise the trigger efficiency. We present the first prototype for the DFEI algorithm, that leverages the power of Graph Neural Networks (GNN). This paper describes the design and development of the algorithm, and its performance in Upgrade I simulated conditions.</p>","PeriodicalId":36026,"journal":{"name":"Computing and Software for Big Science","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10656322/pdf/","citationCount":"0","resultStr":"{\"title\":\"GNN for Deep Full Event Interpretation and Hierarchical Reconstruction of Heavy-Hadron Decays in Proton-Proton Collisions.\",\"authors\":\"Julián García Pardiñas, Marta Calvi, Jonas Eschle, Andrea Mauri, Simone Meloni, Martina Mozzanica, Nicola Serra\",\"doi\":\"10.1007/s41781-023-00107-8\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>The LHCb experiment at the Large Hadron Collider (LHC) is designed to perform high-precision measurements of heavy-hadron decays, which requires the collection of large data samples and a good understanding and suppression of multiple background sources. Both factors are challenged by a fivefold increase in the average number of proton-proton collisions per bunch crossing, corresponding to a change in the detector operation conditions for the LHCb Upgrade I phase, recently started. A further tenfold increase is expected in the Upgrade II phase, planned for the next decade. The limits in the storage capacity of the trigger will bring an inverse relationship between the number of particles selected to be stored per event and the number of events that can be recorded. In addition the background levels will rise due to the enlarged combinatorics. To tackle both challenges, we propose a novel approach, never attempted before in a hadronic collider: a Deep-learning based Full Event Interpretation (DFEI), to perform the simultaneous identification, isolation and hierarchical reconstruction of all the heavy-hadron decay chains per event. This strategy radically contrasts with the standard selection procedure used in LHCb to identify heavy-hadron decays, that looks individually at subsets of particles compatible with being products of specific decay types, disregarding the contextual information from the rest of the event. Following the DFEI approach, once the relevant particles in each event are identified, the rest can be safely removed to optimise the storage space and maximise the trigger efficiency. We present the first prototype for the DFEI algorithm, that leverages the power of Graph Neural Networks (GNN). This paper describes the design and development of the algorithm, and its performance in Upgrade I simulated conditions.</p>\",\"PeriodicalId\":36026,\"journal\":{\"name\":\"Computing and Software for Big Science\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10656322/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computing and Software for Big Science\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1007/s41781-023-00107-8\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2023/11/17 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q1\",\"JCRName\":\"Computer Science\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computing and Software for Big Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s41781-023-00107-8","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2023/11/17 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"Computer Science","Score":null,"Total":0}
引用次数: 0

摘要

大型强子对撞机(LHC)上的大型强子对撞机(LHCb)实验旨在对重强子衰变进行高精度测量,这需要收集大量数据样本,并对多个背景源有很好的理解和抑制。每束交叉的质子-质子碰撞的平均次数增加了五倍,这两个因素都受到了挑战,这与最近开始的LHCb升级I阶段探测器操作条件的变化相对应。预计在计划为下一个十年进行的第二次升级阶段将进一步增加十倍。触发器存储容量的限制将在每个事件选择存储的粒子数量与可以记录的事件数量之间产生反比关系。此外,背景水平将上升,由于扩大的组合。为了解决这两个挑战,我们提出了一种从未在强子对撞机中尝试过的新方法:基于深度学习的全事件解释(DFEI),以执行每个事件中所有重强子衰变链的同时识别,隔离和分层重建。这种策略与LHCb中用于识别重强子衰变的标准选择程序截然不同,后者单独观察与特定衰变类型相容的粒子子集,而忽略事件其余部分的上下文信息。根据DFEI方法,一旦确定了每个事件中的相关粒子,就可以安全地移除其余的粒子,以优化存储空间并最大限度地提高触发效率。我们提出了DFEI算法的第一个原型,它利用了图神经网络(GNN)的力量。本文介绍了该算法的设计与开发,以及该算法在升级I仿真条件下的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
GNN for Deep Full Event Interpretation and Hierarchical Reconstruction of Heavy-Hadron Decays in Proton-Proton Collisions.

The LHCb experiment at the Large Hadron Collider (LHC) is designed to perform high-precision measurements of heavy-hadron decays, which requires the collection of large data samples and a good understanding and suppression of multiple background sources. Both factors are challenged by a fivefold increase in the average number of proton-proton collisions per bunch crossing, corresponding to a change in the detector operation conditions for the LHCb Upgrade I phase, recently started. A further tenfold increase is expected in the Upgrade II phase, planned for the next decade. The limits in the storage capacity of the trigger will bring an inverse relationship between the number of particles selected to be stored per event and the number of events that can be recorded. In addition the background levels will rise due to the enlarged combinatorics. To tackle both challenges, we propose a novel approach, never attempted before in a hadronic collider: a Deep-learning based Full Event Interpretation (DFEI), to perform the simultaneous identification, isolation and hierarchical reconstruction of all the heavy-hadron decay chains per event. This strategy radically contrasts with the standard selection procedure used in LHCb to identify heavy-hadron decays, that looks individually at subsets of particles compatible with being products of specific decay types, disregarding the contextual information from the rest of the event. Following the DFEI approach, once the relevant particles in each event are identified, the rest can be safely removed to optimise the storage space and maximise the trigger efficiency. We present the first prototype for the DFEI algorithm, that leverages the power of Graph Neural Networks (GNN). This paper describes the design and development of the algorithm, and its performance in Upgrade I simulated conditions.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Computing and Software for Big Science
Computing and Software for Big Science Computer Science-Computer Science (miscellaneous)
CiteScore
6.20
自引率
0.00%
发文量
15
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信