{"title":"Quantum image dataset transform (QIDT) for image processing","authors":"Zorkan Erkan, Javad Rahebi, Aref Yelghi","doi":"10.1007/s11128-025-04754-1","DOIUrl":null,"url":null,"abstract":"<div><p>Existing CNN architectures have been developed to train digital image datasets obtained from hardware systems operating with classical bits, such as optical cameras. With the increase of quantum computing algorithms and quantum system providers, academic research is being conducted to combine the strengths of classical computing and quantum algorithms. This fusion allows for the development of hybrid quantum systems, with proposed methods specifically for the quantum representation of digital images. While methods for transforming digital images into quantum-compatible circuits have been proposed, no study has been found on the quantum transformation of entire datasets, especially for the use of fully classical CNN architectures. This article presents the quantum image dataset transform method, which utilizes quantum circuits to transform digital images and create a new dataset of the transformed images. Each of the 10,000 digital images of 28 <span>\\(\\times \\)</span> 28 dimensions in the MNIST handwritten digits dataset is individually sub-parts, and the common weight values of each segment are determined as the phase value to be used in the quantum circuit. The quantum outputs of each sub-part are converted into classical equivalents by creating a quantum converter, and a new digital image is obtained by combining all the sub-parts. The newly generated digital images are labeled as <span>\\({\\textbf {MNIST}} {\\textbf {Q}}^{{\\textbf {+}}}_{{{\\textbf {image}}}}\\)</span> and are publicly shared along with the original MNIST dataset. The paper evaluates both a custom 3-layer CNN architecture and several pre-trained models, including EfficientNetV2B3, ResNet-50, DenseNet-121, and ConvNeXt Tiny. After training for 30 epochs, the 3-layer CNN architecture achieved the highest accuracy of 99.23%, significantly outperforming the pre-trained models, with DenseNet-121 achieving 81.70%, EfficientNetV2B3 64.23%, ResNet-50 53.25%, and ConvNeXt Tiny 53.41%. The results highlight the superior performance of the 3-layer CNN in adapting to the quantum-transformed dataset and demonstrate the potential of quantum transformations to enhance the learning ability of classical CNN models. This foundational research aims to pave the way for further exploration into the integration of quantum-transformed datasets in classical deep learning frameworks.</p></div>","PeriodicalId":746,"journal":{"name":"Quantum Information Processing","volume":"24 6","pages":""},"PeriodicalIF":2.2000,"publicationDate":"2025-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s11128-025-04754-1.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Quantum Information Processing","FirstCategoryId":"101","ListUrlMain":"https://link.springer.com/article/10.1007/s11128-025-04754-1","RegionNum":3,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PHYSICS, MATHEMATICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Existing CNN architectures have been developed to train digital image datasets obtained from hardware systems operating with classical bits, such as optical cameras. With the increase of quantum computing algorithms and quantum system providers, academic research is being conducted to combine the strengths of classical computing and quantum algorithms. This fusion allows for the development of hybrid quantum systems, with proposed methods specifically for the quantum representation of digital images. While methods for transforming digital images into quantum-compatible circuits have been proposed, no study has been found on the quantum transformation of entire datasets, especially for the use of fully classical CNN architectures. This article presents the quantum image dataset transform method, which utilizes quantum circuits to transform digital images and create a new dataset of the transformed images. Each of the 10,000 digital images of 28 \(\times \) 28 dimensions in the MNIST handwritten digits dataset is individually sub-parts, and the common weight values of each segment are determined as the phase value to be used in the quantum circuit. The quantum outputs of each sub-part are converted into classical equivalents by creating a quantum converter, and a new digital image is obtained by combining all the sub-parts. The newly generated digital images are labeled as \({\textbf {MNIST}} {\textbf {Q}}^{{\textbf {+}}}_{{{\textbf {image}}}}\) and are publicly shared along with the original MNIST dataset. The paper evaluates both a custom 3-layer CNN architecture and several pre-trained models, including EfficientNetV2B3, ResNet-50, DenseNet-121, and ConvNeXt Tiny. After training for 30 epochs, the 3-layer CNN architecture achieved the highest accuracy of 99.23%, significantly outperforming the pre-trained models, with DenseNet-121 achieving 81.70%, EfficientNetV2B3 64.23%, ResNet-50 53.25%, and ConvNeXt Tiny 53.41%. The results highlight the superior performance of the 3-layer CNN in adapting to the quantum-transformed dataset and demonstrate the potential of quantum transformations to enhance the learning ability of classical CNN models. This foundational research aims to pave the way for further exploration into the integration of quantum-transformed datasets in classical deep learning frameworks.
现有的CNN架构已经开发用于训练从使用经典比特操作的硬件系统(如光学相机)获得的数字图像数据集。随着量子计算算法和量子系统提供商的增加,将经典计算和量子算法的优势结合起来的学术研究正在进行。这种融合允许混合量子系统的发展,并提出了专门用于数字图像的量子表示的方法。虽然已经提出了将数字图像转换为量子兼容电路的方法,但尚未发现关于整个数据集的量子转换的研究,特别是对于使用完全经典的CNN架构。本文提出了量子图像数据集变换方法,利用量子电路对数字图像进行变换,并将变换后的图像创建一个新的数据集。MNIST手写数字数据集中28个\(\times \) 28维的10000张数字图像中的每一张都是单独的子部分,每个部分的共同权重值被确定为量子电路中使用的相位值。通过创建量子转换器将每个子部分的量子输出转换为经典等效,并将所有子部分组合在一起获得新的数字图像。新生成的数字图像被标记为\({\textbf {MNIST}} {\textbf {Q}}^{{\textbf {+}}}_{{{\textbf {image}}}}\),并与原始MNIST数据集一起公开共享。本文评估了自定义的3层CNN架构和几个预训练模型,包括EfficientNetV2B3、ResNet-50、DenseNet-121和ConvNeXt Tiny。经过30次epoch的训练,3层CNN架构达到了99.23的最高准确率%, significantly outperforming the pre-trained models, with DenseNet-121 achieving 81.70%, EfficientNetV2B3 64.23%, ResNet-50 53.25%, and ConvNeXt Tiny 53.41%. The results highlight the superior performance of the 3-layer CNN in adapting to the quantum-transformed dataset and demonstrate the potential of quantum transformations to enhance the learning ability of classical CNN models. This foundational research aims to pave the way for further exploration into the integration of quantum-transformed datasets in classical deep learning frameworks.
期刊介绍:
Quantum Information Processing is a high-impact, international journal publishing cutting-edge experimental and theoretical research in all areas of Quantum Information Science. Topics of interest include quantum cryptography and communications, entanglement and discord, quantum algorithms, quantum error correction and fault tolerance, quantum computer science, quantum imaging and sensing, and experimental platforms for quantum information. Quantum Information Processing supports and inspires research by providing a comprehensive peer review process, and broadcasting high quality results in a range of formats. These include original papers, letters, broadly focused perspectives, comprehensive review articles, book reviews, and special topical issues. The journal is particularly interested in papers detailing and demonstrating quantum information protocols for cryptography, communications, computation, and sensing.