Modular GAN: positron emission tomography image reconstruction using two generative adversarial networks.

Frontiers in radiology Pub Date : 2024-08-29 eCollection Date: 2024-01-01 DOI:10.3389/fradi.2024.1466498
Rajat Vashistha, Viktor Vegh, Hamed Moradi, Amanda Hammond, Kieran O'Brien, David Reutens
{"title":"Modular GAN: positron emission tomography image reconstruction using two generative adversarial networks.","authors":"Rajat Vashistha, Viktor Vegh, Hamed Moradi, Amanda Hammond, Kieran O'Brien, David Reutens","doi":"10.3389/fradi.2024.1466498","DOIUrl":null,"url":null,"abstract":"<p><strong>Introduction: </strong>The reconstruction of PET images involves converting sinograms, which represent the measured counts of radioactive emissions using detector rings encircling the patient, into meaningful images. However, the quality of PET data acquisition is impacted by physical factors, photon count statistics and detector characteristics, which affect the signal-to-noise ratio, resolution and quantitative accuracy of the resulting images. To address these influences, correction methods have been developed to mitigate each of these issues separately. Recently, generative adversarial networks (GANs) based on machine learning have shown promise in learning the complex mapping between acquired PET data and reconstructed tomographic images. This study aims to investigate the properties of training images that contribute to GAN performance when non-clinical images are used for training. Additionally, we describe a method to correct common PET imaging artefacts without relying on patient-specific anatomical images.</p><p><strong>Methods: </strong>The modular GAN framework includes two GANs. Module 1, resembling Pix2pix architecture, is trained on non-clinical sinogram-image pairs. Training data are optimised by considering image properties defined by metrics. The second module utilises adaptive instance normalisation and style embedding to enhance the quality of images from Module 1. Additional perceptual and patch-based loss functions are employed in training both modules. The performance of the new framework was compared with that of existing methods, (filtered backprojection (FBP) and ordered subset expectation maximisation (OSEM) without and with point spread function (OSEM-PSF)) with respect to correction for attenuation, patient motion and noise in simulated, NEMA phantom and human imaging data. Evaluation metrics included structural similarity (SSIM), peak-signal-to-noise ratio (PSNR), relative root mean squared error (rRMSE) for simulated data, and contrast-to-noise ratio (CNR) for NEMA phantom and human data.</p><p><strong>Results: </strong>For simulated test data, the performance of the proposed framework was both qualitatively and quantitatively superior to that of FBP and OSEM. In the presence of noise, Module 1 generated images with a SSIM of 0.48 and higher. These images exhibited coarse structures that were subsequently refined by Module 2, yielding images with an SSIM higher than 0.71 (at least 22% higher than OSEM). The proposed method was robust against noise and motion. For NEMA phantoms, it achieved higher CNR values than OSEM. For human images, the CNR in brain regions was significantly higher than that of FBP and OSEM (<i>p</i> < 0.05, paired <i>t</i>-test). The CNR of images reconstructed with OSEM-PSF was similar to those reconstructed using the proposed method.</p><p><strong>Conclusion: </strong>The proposed image reconstruction method can produce PET images with artefact correction.</p>","PeriodicalId":73101,"journal":{"name":"Frontiers in radiology","volume":"4 ","pages":"1466498"},"PeriodicalIF":0.0000,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11425657/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in radiology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/fradi.2024.1466498","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/1/1 0:00:00","PubModel":"eCollection","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Introduction: The reconstruction of PET images involves converting sinograms, which represent the measured counts of radioactive emissions using detector rings encircling the patient, into meaningful images. However, the quality of PET data acquisition is impacted by physical factors, photon count statistics and detector characteristics, which affect the signal-to-noise ratio, resolution and quantitative accuracy of the resulting images. To address these influences, correction methods have been developed to mitigate each of these issues separately. Recently, generative adversarial networks (GANs) based on machine learning have shown promise in learning the complex mapping between acquired PET data and reconstructed tomographic images. This study aims to investigate the properties of training images that contribute to GAN performance when non-clinical images are used for training. Additionally, we describe a method to correct common PET imaging artefacts without relying on patient-specific anatomical images.

Methods: The modular GAN framework includes two GANs. Module 1, resembling Pix2pix architecture, is trained on non-clinical sinogram-image pairs. Training data are optimised by considering image properties defined by metrics. The second module utilises adaptive instance normalisation and style embedding to enhance the quality of images from Module 1. Additional perceptual and patch-based loss functions are employed in training both modules. The performance of the new framework was compared with that of existing methods, (filtered backprojection (FBP) and ordered subset expectation maximisation (OSEM) without and with point spread function (OSEM-PSF)) with respect to correction for attenuation, patient motion and noise in simulated, NEMA phantom and human imaging data. Evaluation metrics included structural similarity (SSIM), peak-signal-to-noise ratio (PSNR), relative root mean squared error (rRMSE) for simulated data, and contrast-to-noise ratio (CNR) for NEMA phantom and human data.

Results: For simulated test data, the performance of the proposed framework was both qualitatively and quantitatively superior to that of FBP and OSEM. In the presence of noise, Module 1 generated images with a SSIM of 0.48 and higher. These images exhibited coarse structures that were subsequently refined by Module 2, yielding images with an SSIM higher than 0.71 (at least 22% higher than OSEM). The proposed method was robust against noise and motion. For NEMA phantoms, it achieved higher CNR values than OSEM. For human images, the CNR in brain regions was significantly higher than that of FBP and OSEM (p < 0.05, paired t-test). The CNR of images reconstructed with OSEM-PSF was similar to those reconstructed using the proposed method.

Conclusion: The proposed image reconstruction method can produce PET images with artefact correction.

模块化 GAN:使用两个生成对抗网络进行正电子发射断层图像重建。
简介正电子发射计算机断层显像图像的重建工作包括将正弦曲线图转换成有意义的图像,正弦曲线图代表利用环绕病人的探测器测量到的放射性发射计数。然而,PET 数据采集的质量受到物理因素、光子计数统计和探测器特性的影响,这些因素会影响所生成图像的信噪比、分辨率和定量准确性。为了解决这些影响,人们开发了校正方法来分别缓解这些问题。最近,基于机器学习的生成对抗网络(GANs)在学习获取的 PET 数据和重建的断层图像之间的复杂映射方面显示出良好的前景。本研究旨在研究在使用非临床图像进行训练时,有助于提高 GAN 性能的训练图像属性。此外,我们还介绍了一种无需依赖患者特定解剖图像即可纠正常见 PET 成像伪影的方法:模块化 GAN 框架包括两个 GAN。模块 1 类似 Pix2pix 架构,在非临床正弦图像对上进行训练。训练数据根据指标定义的图像属性进行优化。第二个模块利用自适应实例归一化和风格嵌入来提高模块 1 的图像质量。在训练这两个模块时,还采用了额外的感知损失函数和基于斑块的损失函数。新框架的性能与现有方法(滤波后投影(FBP)和有序子集期望最大化(OSEM),无点扩散函数(OSEM-PSF))进行了比较,以校正模拟、NEMA 模型和人体成像数据中的衰减、患者运动和噪声。评估指标包括结构相似性(SSIM)、峰值信噪比(PSNR)、模拟数据的相对均方根误差(rRMSE),以及 NEMA 人体模型和人体数据的对比信噪比(CNR):对于模拟测试数据,所提出的框架在质量和数量上都优于 FBP 和 OSEM。在存在噪声的情况下,模块 1 生成的图像 SSIM 为 0.48 或更高。这些图像显示出粗略的结构,随后由模块 2 进行细化,生成的图像 SSIM 高于 0.71(比 OSEM 至少高出 22%)。所提出的方法对噪声和运动具有鲁棒性。对于 NEMA 模型,它的 CNR 值高于 OSEM。对于人体图像,大脑区域的 CNR 值明显高于 FBP 和 OSEM(P t 检验)。使用 OSEM-PSF 重建的图像的 CNR 与使用提出的方法重建的图像相似:结论:所提出的图像重建方法可以生成具有伪影校正功能的 PET 图像。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
1.20
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信