Prototype of Machine Learning “as a Service” for CMS Physics in Signal vs Background discrimination

L. Giommi, D. Bonacorsi, V. Kuznetsov
{"title":"Prototype of Machine Learning “as a Service” for CMS Physics in Signal vs Background discrimination","authors":"L. Giommi, D. Bonacorsi, V. Kuznetsov","doi":"10.22323/1.321.0093","DOIUrl":null,"url":null,"abstract":"Big volumes of data are collected and analyzed by LHC experiments at CERN. The success of this scientific challenges is ensured by a great amount of computing power and storage capacity, operated over high performance networks, in very complex LHC computing models on the LHC Computing Grid infrastructure. Now in Run-2 data taking, LHC has an ambitious and broad experimental programme for the coming decades: it includes large investments in detector hardware, and similarly it requires commensurate investment in the R\\&D in software and computing to acquire, manage, process and analyze the shear amounts of data to be recorded in the High-Luminosity LHC (HL-LHC) era. \n \nThe new rise of Artificial Intelligence - related to the current Big Data era, to the technological progress and to a bump in resources democratization and efficient allocation at affordable costs through cloud solutions - is posing new challenges but also offering extremely promising techniques, not only for the commercial world but also for scientific enterprises such as HEP experiments. Machine Learning and Deep Learning are rapidly evolving approaches to characterising and describing data with the potential to radically change how data is reduced and analyzed, also at LHC. \n \nThis work aims at contributing to the construction of a Machine Learning ``as a service'' solution for CMS Physics needs, namely an end-to-end data-service to serve Machine Learning trained model to the CMS software framework. To this ambitious goal, this work contributes firstly with a proof of concept of a first prototype of such infrastructure, and secondly with a specific physics use-case: the Signal versus Background discrimination in the study of CMS all-hadronic top quark decays, done with scalable Machine Learning techniques.","PeriodicalId":346295,"journal":{"name":"Proceedings of Sixth Annual Conference on Large Hadron Collider Physics — PoS(LHCP2018)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of Sixth Annual Conference on Large Hadron Collider Physics — PoS(LHCP2018)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.22323/1.321.0093","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Big volumes of data are collected and analyzed by LHC experiments at CERN. The success of this scientific challenges is ensured by a great amount of computing power and storage capacity, operated over high performance networks, in very complex LHC computing models on the LHC Computing Grid infrastructure. Now in Run-2 data taking, LHC has an ambitious and broad experimental programme for the coming decades: it includes large investments in detector hardware, and similarly it requires commensurate investment in the R\&D in software and computing to acquire, manage, process and analyze the shear amounts of data to be recorded in the High-Luminosity LHC (HL-LHC) era. The new rise of Artificial Intelligence - related to the current Big Data era, to the technological progress and to a bump in resources democratization and efficient allocation at affordable costs through cloud solutions - is posing new challenges but also offering extremely promising techniques, not only for the commercial world but also for scientific enterprises such as HEP experiments. Machine Learning and Deep Learning are rapidly evolving approaches to characterising and describing data with the potential to radically change how data is reduced and analyzed, also at LHC. This work aims at contributing to the construction of a Machine Learning ``as a service'' solution for CMS Physics needs, namely an end-to-end data-service to serve Machine Learning trained model to the CMS software framework. To this ambitious goal, this work contributes firstly with a proof of concept of a first prototype of such infrastructure, and secondly with a specific physics use-case: the Signal versus Background discrimination in the study of CMS all-hadronic top quark decays, done with scalable Machine Learning techniques.
CMS物理在信号与背景识别中的机器学习“即服务”原型
欧洲核子研究中心的大型强子对撞机实验收集和分析了大量数据。这一科学挑战的成功是由大量的计算能力和存储容量保证的,在高性能网络上运行,在LHC计算网格基础设施上的非常复杂的LHC计算模型中。现在在Run-2数据采集中,LHC在未来几十年有一个雄心勃勃的广泛的实验计划:它包括在探测器硬件上的大量投资,同样它需要在软件和计算方面的研发上进行相应的投资,以获取、管理、处理和分析高亮度LHC (HL-LHC)时代将要记录的大量数据。人工智能的新崛起——与当前的大数据时代、技术进步以及通过云解决方案实现资源民主化和高效分配的飞跃有关——不仅为商业世界带来了新的挑战,也为HEP实验等科学企业提供了极具前景的技术。机器学习和深度学习是表征和描述数据的快速发展方法,有可能从根本上改变数据的减少和分析方式,在大型强子对撞机也是如此。本工作旨在为CMS物理需求构建机器学习“即服务”解决方案,即端到端数据服务,为CMS软件框架提供机器学习训练模型。为了实现这一雄心勃勃的目标,这项工作首先证明了这种基础设施的第一个原型的概念,其次是一个特定的物理用例:CMS全强子顶夸克衰变研究中的信号与背景区分,使用可扩展的机器学习技术完成。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信