Simulation of positron emission tomography (PET) images is critical in dynamic imaging protocol optimization, quantitative imaging metric development, deep learning applications, and virtual imaging trials. These applications rely heavily on large volumes of simulated PET data. However, the current state-of-the-art PET image simulation platform is time-prohibitive and computationally intensive. Although deep learning-based generative models have been widely applied to generate PET images, they often fail to adequately account for the differing acquisition times of PET images.
This study seeks to develop and validate a novel deep learning-based method, the noise-aware system generative model (NASGM), to simulate PET images of different acquisition times.
NASGM is based on the conditional generative adversarial network and features a novel dual-domain discriminator that contains a spatial and a frequency branch to leverage information from both domains. A transformer-based structure is applied for the frequency discriminator because of its ability to encode positional information and capture global dependencies. The study is conducted on a simulated dataset, with a public PET/CT dataset as the input activity and attenuation maps, and an analytical PET simulation tool to simulate PET images of different acquisition times. Ablation studies are carried out to confirm the necessity of adopting the dual-domain discriminator. A comprehensive suite of evaluations, including image fidelity assessment, noise measurement, quantitative accuracy validation, task-based assessment, texture analysis, and human observer study, is performed to confirm the realism of generated images. The Wilcoxon signed-rank test with Bonferroni correction is applied to compare the NASGM with other networks in the ablation study at an adjusted p-value , and the alignment of features between the generated and target images is measured by the concordance correlation coefficient (CCC).
The quantitative accuracy measured by the correlation of mean recovery coefficients of tumor groups, and the NASGM-generated images achieved CCC values of 0.95 across most of the acquisition times. This also illustrates NASGM's ability to replicate the partial volume effect in target images. Furthermore, NASGM was also demonstrated to generate images that exhibit noise characteristics and textures closely matching those in the target PET images. In a tumor detection task-based observer study, the synthesized images achieved comparable performance to target images in the clinically relevant task. In the two-alternative forced choice human observer study, human observers achieved an accuracy of ∼50% for all tested acquisition times, confirming that the synthesized and target images are visually indistinguishable for human observers. Moreover, NASGM displayed strong generalizability within the training range, successfully generating images of frame durations not included in the training dataset.
NASGM is developed and validated as a deep learning-based PET simulation framework that offers computationally efficient image generation compared to traditional methods, making it an ideal tool for producing large volumes of simulated PET image datasets across varying acquisition times. Furthermore, the dual-domain discriminator enhances the quality of generated images, while the noise-aware mechanism introduces realistic, controllable noise variability.