{"title":"通过CNN-Transformer组合显微镜图像增强网络的自动活细胞评估","authors":"Wenyuan Chen;Haocong Song;Zhuoran Zhang;Changsheng Dai;Guanqiao Shan;Hang Liu;Aojun Jiang;Chen Sun;Wenkun Dou;Changhai Ru;Clifford Librach;Yu Sun","doi":"10.1109/TASE.2025.3585728","DOIUrl":null,"url":null,"abstract":"Automated morphological measurement of cellular and subcellular structures in live cells is important for evaluating cell functions. Due to their small size and transparent appearance, visualizing cellular and subcellular structures often requires high magnification microscopy and fluorescent staining. However, high magnification microscopy gives a limited field of view, and fluorescent staining alters cell viability and/or activity. Therefore, microscopy image enhancement methods have been developed to predict detailed intracellular structures in live cells. Existing image enhancement networks are mostly CNN-based models lacking global information or Transformer-based models lacking local information. For these purposes, a novel CNN-Transformer combined bilateral U-Net (CTBUnet) is proposed to effectively aggregate both local and global information. Experiments on the collected sperm cell enhancement dataset demonstrate the effectiveness of proposed network for both super-resolution and virtual staining prediction. Note to Practitioners—Automated and accurate intracellular morphology measurement is crucial for cell quality analysis. Microscopy image enhancement methods including super-resolution and virtual staining prediction were proposed to enhance or highlight details of subcellular structures without high magnification microscopy or invasive staining. To effectively combine local and global information, a novel CNN-Transformer combined image enhancement network is proposed. Different from traditional CNN-Transformer combined structures that only directionally fuse outputs from CNN and Transformer, the proposed bilateral fusion module bidirectionally fuses and exchanges features from CNN and Transformer. The proposed bilateral fusion module incorporates not only channel-wise fusion but also spatial-wise fusion to effectively aggregate local and global features. Additionally, a region-aware attention gate is proposed to urge the network to only focus on reconstructing cell structures regardless of background. The proposed method outperformed existing networks with a better enhancement effect for subcellular details.","PeriodicalId":51060,"journal":{"name":"IEEE Transactions on Automation Science and Engineering","volume":"22 ","pages":"18269-18280"},"PeriodicalIF":6.4000,"publicationDate":"2025-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Automated Live Cell Evaluation via a CNN-Transformer Combined Microscopy Image Enhancement Network\",\"authors\":\"Wenyuan Chen;Haocong Song;Zhuoran Zhang;Changsheng Dai;Guanqiao Shan;Hang Liu;Aojun Jiang;Chen Sun;Wenkun Dou;Changhai Ru;Clifford Librach;Yu Sun\",\"doi\":\"10.1109/TASE.2025.3585728\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Automated morphological measurement of cellular and subcellular structures in live cells is important for evaluating cell functions. Due to their small size and transparent appearance, visualizing cellular and subcellular structures often requires high magnification microscopy and fluorescent staining. However, high magnification microscopy gives a limited field of view, and fluorescent staining alters cell viability and/or activity. Therefore, microscopy image enhancement methods have been developed to predict detailed intracellular structures in live cells. Existing image enhancement networks are mostly CNN-based models lacking global information or Transformer-based models lacking local information. For these purposes, a novel CNN-Transformer combined bilateral U-Net (CTBUnet) is proposed to effectively aggregate both local and global information. Experiments on the collected sperm cell enhancement dataset demonstrate the effectiveness of proposed network for both super-resolution and virtual staining prediction. Note to Practitioners—Automated and accurate intracellular morphology measurement is crucial for cell quality analysis. Microscopy image enhancement methods including super-resolution and virtual staining prediction were proposed to enhance or highlight details of subcellular structures without high magnification microscopy or invasive staining. To effectively combine local and global information, a novel CNN-Transformer combined image enhancement network is proposed. Different from traditional CNN-Transformer combined structures that only directionally fuse outputs from CNN and Transformer, the proposed bilateral fusion module bidirectionally fuses and exchanges features from CNN and Transformer. The proposed bilateral fusion module incorporates not only channel-wise fusion but also spatial-wise fusion to effectively aggregate local and global features. Additionally, a region-aware attention gate is proposed to urge the network to only focus on reconstructing cell structures regardless of background. The proposed method outperformed existing networks with a better enhancement effect for subcellular details.\",\"PeriodicalId\":51060,\"journal\":{\"name\":\"IEEE Transactions on Automation Science and Engineering\",\"volume\":\"22 \",\"pages\":\"18269-18280\"},\"PeriodicalIF\":6.4000,\"publicationDate\":\"2025-07-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Automation Science and Engineering\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/11068996/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automation Science and Engineering","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/11068996/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Automated Live Cell Evaluation via a CNN-Transformer Combined Microscopy Image Enhancement Network
Automated morphological measurement of cellular and subcellular structures in live cells is important for evaluating cell functions. Due to their small size and transparent appearance, visualizing cellular and subcellular structures often requires high magnification microscopy and fluorescent staining. However, high magnification microscopy gives a limited field of view, and fluorescent staining alters cell viability and/or activity. Therefore, microscopy image enhancement methods have been developed to predict detailed intracellular structures in live cells. Existing image enhancement networks are mostly CNN-based models lacking global information or Transformer-based models lacking local information. For these purposes, a novel CNN-Transformer combined bilateral U-Net (CTBUnet) is proposed to effectively aggregate both local and global information. Experiments on the collected sperm cell enhancement dataset demonstrate the effectiveness of proposed network for both super-resolution and virtual staining prediction. Note to Practitioners—Automated and accurate intracellular morphology measurement is crucial for cell quality analysis. Microscopy image enhancement methods including super-resolution and virtual staining prediction were proposed to enhance or highlight details of subcellular structures without high magnification microscopy or invasive staining. To effectively combine local and global information, a novel CNN-Transformer combined image enhancement network is proposed. Different from traditional CNN-Transformer combined structures that only directionally fuse outputs from CNN and Transformer, the proposed bilateral fusion module bidirectionally fuses and exchanges features from CNN and Transformer. The proposed bilateral fusion module incorporates not only channel-wise fusion but also spatial-wise fusion to effectively aggregate local and global features. Additionally, a region-aware attention gate is proposed to urge the network to only focus on reconstructing cell structures regardless of background. The proposed method outperformed existing networks with a better enhancement effect for subcellular details.
期刊介绍:
The IEEE Transactions on Automation Science and Engineering (T-ASE) publishes fundamental papers on Automation, emphasizing scientific results that advance efficiency, quality, productivity, and reliability. T-ASE encourages interdisciplinary approaches from computer science, control systems, electrical engineering, mathematics, mechanical engineering, operations research, and other fields. T-ASE welcomes results relevant to industries such as agriculture, biotechnology, healthcare, home automation, maintenance, manufacturing, pharmaceuticals, retail, security, service, supply chains, and transportation. T-ASE addresses a research community willing to integrate knowledge across disciplines and industries. For this purpose, each paper includes a Note to Practitioners that summarizes how its results can be applied or how they might be extended to apply in practice.