核外感知多gpu渲染大规模场景可视化

IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS
Milan Jaros, Lubomir Riha, Petr Strakos, Tomas Kozubek
{"title":"核外感知多gpu渲染大规模场景可视化","authors":"Milan Jaros,&nbsp;Lubomir Riha,&nbsp;Petr Strakos,&nbsp;Tomas Kozubek","doi":"10.1016/j.future.2025.108109","DOIUrl":null,"url":null,"abstract":"<div><div>We introduce an out-of-core method for multi-GPU path tracing of large-scale scenes, leveraging memory access analysis to optimize data distribution. Our approach enables efficient rendering on multi-GPU systems even when the combined GPU memory is smaller than the total scene size. By partitioning the scene between GPU and CPU memory at the memory management level, our method strategically distributes or replicates scene data across GPUs while storing the remaining data in CPU memory. This hybrid memory strategy ensures efficient access patterns and minimizes performance bottlenecks, facilitating high-quality rendering of massive scenes. The main contribution of this rendering approach is that it enables GPUs to perform accelerated path-tracing for massive scenes that would otherwise be rendered only by CPUs. Previous work on this topic has enabled the rendering of massive scenes that are distributed among memories of all GPUs on one server. However, in that method, it is not possible to render scenes larger than the total size of all GPU memories. In our paper, we show how to break this barrier and how to combine the capacity of GPU memories with CPU main memory to render massive scenes with only a minor impact on the performance. When applied to a small system with 4xGPUs, the results are equivalent to a more powerful system with 16xGPUs if using the same number of GPUs as on the smaller system. Therefore, users can use even small systems to render massive scenes.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"175 ","pages":"Article 108109"},"PeriodicalIF":6.2000,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Out-of-core aware multi-GPU rendering for large-scale scene visualization\",\"authors\":\"Milan Jaros,&nbsp;Lubomir Riha,&nbsp;Petr Strakos,&nbsp;Tomas Kozubek\",\"doi\":\"10.1016/j.future.2025.108109\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>We introduce an out-of-core method for multi-GPU path tracing of large-scale scenes, leveraging memory access analysis to optimize data distribution. Our approach enables efficient rendering on multi-GPU systems even when the combined GPU memory is smaller than the total scene size. By partitioning the scene between GPU and CPU memory at the memory management level, our method strategically distributes or replicates scene data across GPUs while storing the remaining data in CPU memory. This hybrid memory strategy ensures efficient access patterns and minimizes performance bottlenecks, facilitating high-quality rendering of massive scenes. The main contribution of this rendering approach is that it enables GPUs to perform accelerated path-tracing for massive scenes that would otherwise be rendered only by CPUs. Previous work on this topic has enabled the rendering of massive scenes that are distributed among memories of all GPUs on one server. However, in that method, it is not possible to render scenes larger than the total size of all GPU memories. In our paper, we show how to break this barrier and how to combine the capacity of GPU memories with CPU main memory to render massive scenes with only a minor impact on the performance. When applied to a small system with 4xGPUs, the results are equivalent to a more powerful system with 16xGPUs if using the same number of GPUs as on the smaller system. Therefore, users can use even small systems to render massive scenes.</div></div>\",\"PeriodicalId\":55132,\"journal\":{\"name\":\"Future Generation Computer Systems-The International Journal of Escience\",\"volume\":\"175 \",\"pages\":\"Article 108109\"},\"PeriodicalIF\":6.2000,\"publicationDate\":\"2025-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Future Generation Computer Systems-The International Journal of Escience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167739X25004030\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X25004030","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0

摘要

我们引入了一种核外方法,用于大规模场景的多gpu路径跟踪,利用内存访问分析来优化数据分布。我们的方法可以在多GPU系统上实现高效的渲染,即使合并的GPU内存小于总场景大小。通过在内存管理级别将场景划分为GPU和CPU内存,我们的方法策略性地在GPU上分布或复制场景数据,同时将剩余数据存储在CPU内存中。这种混合内存策略确保了高效的访问模式,并最大限度地减少了性能瓶颈,促进了大规模场景的高质量渲染。这种渲染方法的主要贡献是,它使gpu能够为大规模场景执行加速路径跟踪,否则只能由cpu渲染。之前关于这个主题的工作已经允许在一台服务器上所有gpu的内存中分布大量场景的渲染。然而,在这种方法中,渲染的场景不可能大于所有GPU内存的总大小。在我们的论文中,我们展示了如何打破这个障碍,以及如何结合GPU内存和CPU主内存的容量来渲染大量场景,而对性能的影响很小。当应用于具有4xgpu的小型系统时,如果使用与小型系统相同数量的gpu,则结果相当于具有16xgpu的更强大的系统。因此,用户甚至可以使用小系统来渲染大规模场景。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Out-of-core aware multi-GPU rendering for large-scale scene visualization
We introduce an out-of-core method for multi-GPU path tracing of large-scale scenes, leveraging memory access analysis to optimize data distribution. Our approach enables efficient rendering on multi-GPU systems even when the combined GPU memory is smaller than the total scene size. By partitioning the scene between GPU and CPU memory at the memory management level, our method strategically distributes or replicates scene data across GPUs while storing the remaining data in CPU memory. This hybrid memory strategy ensures efficient access patterns and minimizes performance bottlenecks, facilitating high-quality rendering of massive scenes. The main contribution of this rendering approach is that it enables GPUs to perform accelerated path-tracing for massive scenes that would otherwise be rendered only by CPUs. Previous work on this topic has enabled the rendering of massive scenes that are distributed among memories of all GPUs on one server. However, in that method, it is not possible to render scenes larger than the total size of all GPU memories. In our paper, we show how to break this barrier and how to combine the capacity of GPU memories with CPU main memory to render massive scenes with only a minor impact on the performance. When applied to a small system with 4xGPUs, the results are equivalent to a more powerful system with 16xGPUs if using the same number of GPUs as on the smaller system. Therefore, users can use even small systems to render massive scenes.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
19.90
自引率
2.70%
发文量
376
审稿时长
10.6 months
期刊介绍: Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications. Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration. Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信