Lensless Imaging Based on Dual-Input Physics-Driven Neural Network

IF 3.7 Q2 MATERIALS SCIENCE, MULTIDISCIPLINARY
Jiale Zuo, Ju Tang, Mengmeng Zhang, Jiawei Zhang, Zhenbo Ren, Jianglei Di, Jianlin Zhao
{"title":"Lensless Imaging Based on Dual-Input Physics-Driven Neural Network","authors":"Jiale Zuo,&nbsp;Ju Tang,&nbsp;Mengmeng Zhang,&nbsp;Jiawei Zhang,&nbsp;Zhenbo Ren,&nbsp;Jianglei Di,&nbsp;Jianlin Zhao","doi":"10.1002/adpr.202400029","DOIUrl":null,"url":null,"abstract":"<p>Lensless imaging, as a novel computational imaging technique, has attracted great attention due to its simplicity, compactness, and flexibility. This technique analyzes and processes the diffraction of an object to obtain complex amplitude information. However, traditional algorithms such as Gerchberg-Saxton (G–S) algorithm tend to exhibit significant errors in complex amplitude retrieval, particularly for edge information. Additional constraints have to be incorporated on top of amplitude constraints to enhance the accuracy. Recently, deep learning has shown promising results in optical imaging. However, it requires a large amount of training data. To address these issues, a novel approach called dual-input physics-driven network (DPNN) is proposed for lensless imaging. DPNN utilizes two diffractions recorded at different distances as inputs and uses an unsupervised approach that combines physical imaging model to reconstruct object information. DPNN adopts a U-Net 3+ architecture with a loss function of mean absolute error (MAE) to better capture diffraction features. DPNN achieves highly accurate reconstruction without requiring extensive data and being immune to background noise. Based on different diffraction intervals, noise levels, and imaging models, DPNN exhibits superior capabilities in peak signal-to-noise ratio and structural similarity compared with conventional methods, effectively achieving accurate phase or amplitude information reconstruction.</p>","PeriodicalId":7263,"journal":{"name":"Advanced Photonics Research","volume":"5 11","pages":""},"PeriodicalIF":3.7000,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/adpr.202400029","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advanced Photonics Research","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/adpr.202400029","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MATERIALS SCIENCE, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

Abstract

Lensless imaging, as a novel computational imaging technique, has attracted great attention due to its simplicity, compactness, and flexibility. This technique analyzes and processes the diffraction of an object to obtain complex amplitude information. However, traditional algorithms such as Gerchberg-Saxton (G–S) algorithm tend to exhibit significant errors in complex amplitude retrieval, particularly for edge information. Additional constraints have to be incorporated on top of amplitude constraints to enhance the accuracy. Recently, deep learning has shown promising results in optical imaging. However, it requires a large amount of training data. To address these issues, a novel approach called dual-input physics-driven network (DPNN) is proposed for lensless imaging. DPNN utilizes two diffractions recorded at different distances as inputs and uses an unsupervised approach that combines physical imaging model to reconstruct object information. DPNN adopts a U-Net 3+ architecture with a loss function of mean absolute error (MAE) to better capture diffraction features. DPNN achieves highly accurate reconstruction without requiring extensive data and being immune to background noise. Based on different diffraction intervals, noise levels, and imaging models, DPNN exhibits superior capabilities in peak signal-to-noise ratio and structural similarity compared with conventional methods, effectively achieving accurate phase or amplitude information reconstruction.

Abstract Image

基于双输入物理驱动神经网络的无透镜成像技术
无透镜成像是一种新型计算成像技术,因其简单、紧凑和灵活而备受关注。这种技术通过分析和处理物体的衍射来获取复杂的振幅信息。然而,传统算法(如 Gerchberg-Saxton (G-S) 算法)在复杂振幅检索中往往表现出明显的误差,尤其是边缘信息。为了提高准确性,必须在振幅约束的基础上加入额外的约束。最近,深度学习在光学成像方面取得了可喜的成果。然而,它需要大量的训练数据。为了解决这些问题,我们提出了一种用于无透镜成像的名为双输入物理驱动网络(DPNN)的新方法。DPNN 利用在不同距离记录的两个衍射作为输入,并采用无监督的方法,结合物理成像模型来重建物体信息。DPNN 采用 U-Net 3+ 架构,损失函数为平均绝对误差(MAE),能更好地捕捉衍射特征。DPNN 无需大量数据即可实现高精度重建,并且不受背景噪声影响。基于不同的衍射区间、噪声水平和成像模型,DPNN 在峰值信噪比和结构相似性方面都表现出优于传统方法的能力,可有效实现精确的相位或振幅信息重建。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
2.70%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信