Rajat Vashistha, Viktor Vegh, Hamed Moradi, Amanda Hammond, Kieran O'Brien, David Reutens
{"title":"模块化 GAN:使用两个生成对抗网络进行正电子发射断层图像重建。","authors":"Rajat Vashistha, Viktor Vegh, Hamed Moradi, Amanda Hammond, Kieran O'Brien, David Reutens","doi":"10.3389/fradi.2024.1466498","DOIUrl":null,"url":null,"abstract":"<p><strong>Introduction: </strong>The reconstruction of PET images involves converting sinograms, which represent the measured counts of radioactive emissions using detector rings encircling the patient, into meaningful images. However, the quality of PET data acquisition is impacted by physical factors, photon count statistics and detector characteristics, which affect the signal-to-noise ratio, resolution and quantitative accuracy of the resulting images. To address these influences, correction methods have been developed to mitigate each of these issues separately. Recently, generative adversarial networks (GANs) based on machine learning have shown promise in learning the complex mapping between acquired PET data and reconstructed tomographic images. This study aims to investigate the properties of training images that contribute to GAN performance when non-clinical images are used for training. Additionally, we describe a method to correct common PET imaging artefacts without relying on patient-specific anatomical images.</p><p><strong>Methods: </strong>The modular GAN framework includes two GANs. Module 1, resembling Pix2pix architecture, is trained on non-clinical sinogram-image pairs. Training data are optimised by considering image properties defined by metrics. The second module utilises adaptive instance normalisation and style embedding to enhance the quality of images from Module 1. Additional perceptual and patch-based loss functions are employed in training both modules. The performance of the new framework was compared with that of existing methods, (filtered backprojection (FBP) and ordered subset expectation maximisation (OSEM) without and with point spread function (OSEM-PSF)) with respect to correction for attenuation, patient motion and noise in simulated, NEMA phantom and human imaging data. Evaluation metrics included structural similarity (SSIM), peak-signal-to-noise ratio (PSNR), relative root mean squared error (rRMSE) for simulated data, and contrast-to-noise ratio (CNR) for NEMA phantom and human data.</p><p><strong>Results: </strong>For simulated test data, the performance of the proposed framework was both qualitatively and quantitatively superior to that of FBP and OSEM. In the presence of noise, Module 1 generated images with a SSIM of 0.48 and higher. These images exhibited coarse structures that were subsequently refined by Module 2, yielding images with an SSIM higher than 0.71 (at least 22% higher than OSEM). The proposed method was robust against noise and motion. For NEMA phantoms, it achieved higher CNR values than OSEM. For human images, the CNR in brain regions was significantly higher than that of FBP and OSEM (<i>p</i> < 0.05, paired <i>t</i>-test). The CNR of images reconstructed with OSEM-PSF was similar to those reconstructed using the proposed method.</p><p><strong>Conclusion: </strong>The proposed image reconstruction method can produce PET images with artefact correction.</p>","PeriodicalId":73101,"journal":{"name":"Frontiers in radiology","volume":"4 ","pages":"1466498"},"PeriodicalIF":0.0000,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11425657/pdf/","citationCount":"0","resultStr":"{\"title\":\"Modular GAN: positron emission tomography image reconstruction using two generative adversarial networks.\",\"authors\":\"Rajat Vashistha, Viktor Vegh, Hamed Moradi, Amanda Hammond, Kieran O'Brien, David Reutens\",\"doi\":\"10.3389/fradi.2024.1466498\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Introduction: </strong>The reconstruction of PET images involves converting sinograms, which represent the measured counts of radioactive emissions using detector rings encircling the patient, into meaningful images. However, the quality of PET data acquisition is impacted by physical factors, photon count statistics and detector characteristics, which affect the signal-to-noise ratio, resolution and quantitative accuracy of the resulting images. To address these influences, correction methods have been developed to mitigate each of these issues separately. Recently, generative adversarial networks (GANs) based on machine learning have shown promise in learning the complex mapping between acquired PET data and reconstructed tomographic images. This study aims to investigate the properties of training images that contribute to GAN performance when non-clinical images are used for training. Additionally, we describe a method to correct common PET imaging artefacts without relying on patient-specific anatomical images.</p><p><strong>Methods: </strong>The modular GAN framework includes two GANs. Module 1, resembling Pix2pix architecture, is trained on non-clinical sinogram-image pairs. Training data are optimised by considering image properties defined by metrics. The second module utilises adaptive instance normalisation and style embedding to enhance the quality of images from Module 1. Additional perceptual and patch-based loss functions are employed in training both modules. The performance of the new framework was compared with that of existing methods, (filtered backprojection (FBP) and ordered subset expectation maximisation (OSEM) without and with point spread function (OSEM-PSF)) with respect to correction for attenuation, patient motion and noise in simulated, NEMA phantom and human imaging data. Evaluation metrics included structural similarity (SSIM), peak-signal-to-noise ratio (PSNR), relative root mean squared error (rRMSE) for simulated data, and contrast-to-noise ratio (CNR) for NEMA phantom and human data.</p><p><strong>Results: </strong>For simulated test data, the performance of the proposed framework was both qualitatively and quantitatively superior to that of FBP and OSEM. In the presence of noise, Module 1 generated images with a SSIM of 0.48 and higher. These images exhibited coarse structures that were subsequently refined by Module 2, yielding images with an SSIM higher than 0.71 (at least 22% higher than OSEM). The proposed method was robust against noise and motion. For NEMA phantoms, it achieved higher CNR values than OSEM. For human images, the CNR in brain regions was significantly higher than that of FBP and OSEM (<i>p</i> < 0.05, paired <i>t</i>-test). The CNR of images reconstructed with OSEM-PSF was similar to those reconstructed using the proposed method.</p><p><strong>Conclusion: </strong>The proposed image reconstruction method can produce PET images with artefact correction.</p>\",\"PeriodicalId\":73101,\"journal\":{\"name\":\"Frontiers in radiology\",\"volume\":\"4 \",\"pages\":\"1466498\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11425657/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Frontiers in radiology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3389/fradi.2024.1466498\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2024/1/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in radiology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/fradi.2024.1466498","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/1/1 0:00:00","PubModel":"eCollection","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
简介正电子发射计算机断层显像图像的重建工作包括将正弦曲线图转换成有意义的图像,正弦曲线图代表利用环绕病人的探测器测量到的放射性发射计数。然而,PET 数据采集的质量受到物理因素、光子计数统计和探测器特性的影响,这些因素会影响所生成图像的信噪比、分辨率和定量准确性。为了解决这些影响,人们开发了校正方法来分别缓解这些问题。最近,基于机器学习的生成对抗网络(GANs)在学习获取的 PET 数据和重建的断层图像之间的复杂映射方面显示出良好的前景。本研究旨在研究在使用非临床图像进行训练时,有助于提高 GAN 性能的训练图像属性。此外,我们还介绍了一种无需依赖患者特定解剖图像即可纠正常见 PET 成像伪影的方法:模块化 GAN 框架包括两个 GAN。模块 1 类似 Pix2pix 架构,在非临床正弦图像对上进行训练。训练数据根据指标定义的图像属性进行优化。第二个模块利用自适应实例归一化和风格嵌入来提高模块 1 的图像质量。在训练这两个模块时,还采用了额外的感知损失函数和基于斑块的损失函数。新框架的性能与现有方法(滤波后投影(FBP)和有序子集期望最大化(OSEM),无点扩散函数(OSEM-PSF))进行了比较,以校正模拟、NEMA 模型和人体成像数据中的衰减、患者运动和噪声。评估指标包括结构相似性(SSIM)、峰值信噪比(PSNR)、模拟数据的相对均方根误差(rRMSE),以及 NEMA 人体模型和人体数据的对比信噪比(CNR):对于模拟测试数据,所提出的框架在质量和数量上都优于 FBP 和 OSEM。在存在噪声的情况下,模块 1 生成的图像 SSIM 为 0.48 或更高。这些图像显示出粗略的结构,随后由模块 2 进行细化,生成的图像 SSIM 高于 0.71(比 OSEM 至少高出 22%)。所提出的方法对噪声和运动具有鲁棒性。对于 NEMA 模型,它的 CNR 值高于 OSEM。对于人体图像,大脑区域的 CNR 值明显高于 FBP 和 OSEM(P t 检验)。使用 OSEM-PSF 重建的图像的 CNR 与使用提出的方法重建的图像相似:结论:所提出的图像重建方法可以生成具有伪影校正功能的 PET 图像。
Modular GAN: positron emission tomography image reconstruction using two generative adversarial networks.
Introduction: The reconstruction of PET images involves converting sinograms, which represent the measured counts of radioactive emissions using detector rings encircling the patient, into meaningful images. However, the quality of PET data acquisition is impacted by physical factors, photon count statistics and detector characteristics, which affect the signal-to-noise ratio, resolution and quantitative accuracy of the resulting images. To address these influences, correction methods have been developed to mitigate each of these issues separately. Recently, generative adversarial networks (GANs) based on machine learning have shown promise in learning the complex mapping between acquired PET data and reconstructed tomographic images. This study aims to investigate the properties of training images that contribute to GAN performance when non-clinical images are used for training. Additionally, we describe a method to correct common PET imaging artefacts without relying on patient-specific anatomical images.
Methods: The modular GAN framework includes two GANs. Module 1, resembling Pix2pix architecture, is trained on non-clinical sinogram-image pairs. Training data are optimised by considering image properties defined by metrics. The second module utilises adaptive instance normalisation and style embedding to enhance the quality of images from Module 1. Additional perceptual and patch-based loss functions are employed in training both modules. The performance of the new framework was compared with that of existing methods, (filtered backprojection (FBP) and ordered subset expectation maximisation (OSEM) without and with point spread function (OSEM-PSF)) with respect to correction for attenuation, patient motion and noise in simulated, NEMA phantom and human imaging data. Evaluation metrics included structural similarity (SSIM), peak-signal-to-noise ratio (PSNR), relative root mean squared error (rRMSE) for simulated data, and contrast-to-noise ratio (CNR) for NEMA phantom and human data.
Results: For simulated test data, the performance of the proposed framework was both qualitatively and quantitatively superior to that of FBP and OSEM. In the presence of noise, Module 1 generated images with a SSIM of 0.48 and higher. These images exhibited coarse structures that were subsequently refined by Module 2, yielding images with an SSIM higher than 0.71 (at least 22% higher than OSEM). The proposed method was robust against noise and motion. For NEMA phantoms, it achieved higher CNR values than OSEM. For human images, the CNR in brain regions was significantly higher than that of FBP and OSEM (p < 0.05, paired t-test). The CNR of images reconstructed with OSEM-PSF was similar to those reconstructed using the proposed method.
Conclusion: The proposed image reconstruction method can produce PET images with artefact correction.