{"title":"Reconstructing images with attention generative adversarial network against adversarial attacks","authors":"Xiong Shen, Yiqin Lu, Zhe Cheng, Zhongshu Mao, Zhang Yang, Jiancheng Qin","doi":"10.1117/1.jei.33.3.033029","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033029","url":null,"abstract":"Deep learning is widely used in the field of computer vision, but the emergence of adversarial examples threatens its application. How to effectively detect adversarial examples and correct their labels has become a problem to be solved in this application field. Generative adversarial networks (GANs) can effectively learn the features from images. Based on GAN, this work proposes a defense method called “Reconstructing images with GAN” (RIG). The adversarial examples are generated by attack algorithms reconstructed by the trained generator of RIG to eliminate the perturbations of the adversarial examples, which disturb the models for classification, so that the models can restore their labels when classifying the reconstructed images. In addition, to improve the defensive performance of RIG, the attention mechanism (AM) is introduced to enhance the defense effect of RIG, which is called reconstructing images with attention GAN (RIAG). Experiments show that RIG and RIAG can effectively eliminate the perturbations of the adversarial examples. The results also show that RIAG has a better defensive performance than RIG in eliminating the perturbations of adversarial examples, which indicates that the introduction of AM can effectively improve the defense effect of RIG.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"135 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"AFFNet: adversarial feature fusion network for super-resolution image reconstruction in remote sensing images","authors":"Qian Zhao, Qianxi Yin","doi":"10.1117/1.jei.33.3.033032","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033032","url":null,"abstract":"As an important source of Earth surface information, remote sensing image has the problems of rough and fuzzy image details and poor perception quality, which affect further analysis and application of geographic information. To address the above problems, we introduce the adversarial feature fusion network with an attention-based mechanism for super-resolution reconstruction of remote sensing images in this paper. First, residual structures are designed in the generator to enhance the deep feature extraction capability of remote sensing images. The residual structure is composed of the depthwise over-parameterized convolution and self-attention mechanism, which work synergistically to extract deep feature information from remote sensing images. Second, coordinate attention feature fusion module is introduced at the feature fusion stage, which can fuse shallow features and deep high-level features. Therefore, it can enhance the attention of the model to different features and better fuse inconsistent semantic features. Finally, we design the pixel-attention upsampling module in the up-sampling stage. It adaptively focuses on the most information-rich regions of a pixel and restores the image details more accurately. We conducted extensive experiments on several remote sensing image datasets, and the results showed that compared with current advanced models, our method can better restore the details in the image and achieve good subjective visual effects, which also verifies the effectiveness and superiority of the algorithm proposed in this paper.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"36 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Lightweight deep and cross residual skip connection separable CNN for plant leaf diseases classification","authors":"Naresh Vedhamuru, Ramanathan Malmathanraj, Ponnusamy Palanisamy","doi":"10.1117/1.jei.33.3.033035","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033035","url":null,"abstract":"Crop diseases have an adverse effect on the yield, productivity, and quality of agricultural produce, which threatens the safety and security of the global feature of food supply. Addressing and controlling plant diseases through implementation of timely disease management strategies to reduce their transmission are essential for ensuring minimal crop loss, and addressing the increasing demand for food worldwide as the population continues to increase in a steadfast manner. Crop disease mitigation measures involve preventive monitoring, resulting in early detection and classification of plant diseases for effective agricultural procedure to improve crop yield. Early detection and accurate diagnosis of plant diseases enables farmers to deploy disease management strategies, such interventions are critical for better management contributing to higher crop output by curbing the spread of infection and limiting the extent of damage caused by diseases. We propose and implement a deep and cross residual skip connection separable convolutional neural network (DCRSCSCNN) for identifying and classifying leaf diseases for crops including apple, corn, cucumber, grape, potato, and guava. The significant feature of DCRSCSCNN includes residual skip connection and cross residual skip connection separable convolution block. The usage of residual skip connections assists in fixing the gradient vanishing issue faced by network architecture. The employment of separable convolution decreases the number of parameters, which leads to a model with a reduced size. So far, there has been limited exploration or investigation of leveraging separable convolution within lightweight neural networks. Extensive evaluation of several training and test sets using distinct datasets demonstrate that the proposed DCRSCSCNN outperforms other state-of-the-art approaches. The DCRSCSCNN achieved exceptional classification and identification accuracy rates of 99.89% for apple, 98.72% for corn, 100% for cucumber, 99.78% for grape, 100% for potato, 99.69% for guava1, and 99.08% for guava2 datasets.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"74 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Weina Dong, Jia Liu, Lifeng Chen, Wenquan Sun, Xiaozhong Pan
{"title":"Stega4NeRF: cover selection steganography for neural radiance fields","authors":"Weina Dong, Jia Liu, Lifeng Chen, Wenquan Sun, Xiaozhong Pan","doi":"10.1117/1.jei.33.3.033031","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033031","url":null,"abstract":"The implicit neural representation of visual data (such as images, videos, and 3D models) has become a current hotspot in computer vision research. This work proposes a cover selection steganography scheme for neural radiance fields (NeRFs). The message sender first trains an NeRF model selecting any viewpoint in 3D space as the viewpoint key Kv, to generate a unique secret viewpoint image. Subsequently, a message extractor is trained using overfitting to establish a one-to-one mapping between the secret viewpoint image and the secret message. To address the issue of securely transmitting the message extractor in traditional steganography, the message extractor is concealed within a hybrid model performing standard classification tasks. The receiver possesses a shared extractor key Ke, which is used to recover the message extractor from the hybrid model. Then the secret viewpoint image is obtained by NeRF through the viewpoint key Kv, and the secret message is extracted by inputting it into the message extractor. Experimental results demonstrate that the trained message extractor achieves high-speed steganography with a large capacity and attains a 100% message embedding. Additionally, the vast viewpoint key space of NeRF ensures the concealment of the scheme.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"81 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Low-light image enhancement using negative feedback pulse coupled neural network","authors":"Ping Gao, Guidong Zhang, Lingling Chen, Xiaoyun Chen","doi":"10.1117/1.jei.33.3.033037","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033037","url":null,"abstract":"Low-light image enhancement, fundamentally an ill-posed problem, seeks to simultaneously provide superior visual effects and preserve the natural appearance. Current methodologies often exhibit limitations in contrast enhancement, noise reduction, and the mitigation of halo artifacts. Negative feedback pulse coupled neural network (NFPCNN) is proposed to provide a well posed solution based on uniform distribution in contrast enhancement. The negative feedback dynamically adjusts the attenuation amplitude of neuron threshold based on recent neuronal ignited state. Neurons in the concentrated brightness area arrange smaller attenuation amplitude to enhance the local contrast, whereas neurons in the sparse area set larger attenuation amplitude. NFPCNN makes up for the negligence of pulse coupled neural network in the brightness distribution of the input image. Consistent with Weber–Fechner law, gamma correction is employed to adjust the output of NFPCNN. Although contrast enhancement can improve detail expressiveness, it might also introduce artifacts or aggravate noise. To mitigate these issues, the bilateral filter is employed to suppress halo artifacts. Brightness is used as coefficient to refine the Relativity-of-Gaussian noise suppression method. Experimental results show that the proposed method can effectively suppress noise while enhancing image contrast.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"28 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Face antispoofing method based on single-modal and lightweight network","authors":"Guoxiang Tong, Xinrong Yan","doi":"10.1117/1.jei.33.3.033030","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033030","url":null,"abstract":"In the field of face antispoofing, researchers are increasingly focusing their efforts on multimodal and feature fusion. While multimodal approaches are more effective than single-modal ones, they often come with a huge number of parameters, require significant computational resources, and pose challenges for execution on mobile devices. To address the real-time problem, we propose a fast and lightweight framework based on ShuffleNet V2. Our approach takes patch-level images as input, enhances unit performance by introducing an attention module, and addresses dataset sample imbalance issues through the focal loss function. The framework effectively tackles the real-time constraints of the model. We evaluate the performance of our model on CASIA-FASD, Replay-Attack, and MSU-MFSD datasets. The results demonstrate that our method outperforms the current state-of-the-art methods in both intratest and intertest scenarios. Furthermore, our network has only 0.84 M parameters and 0.81 GFlops, making it suitable for deployment in mobile and real-time settings. Our work can serve as a valuable reference for researchers seeking to develop single-modal face antispoofing methods suitable for mobile and real-time applications.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"230 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Adaptive sparse attention module based on reciprocal nearest neighbors","authors":"Zhonggui Sun, Can Zhang, Mingzhu Zhang","doi":"10.1117/1.jei.33.3.033038","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033038","url":null,"abstract":"The attention mechanism has become a crucial technique in deep feature representation for computer vision tasks. Using a similarity matrix, it enhances the current feature point with global context from the feature map of the network. However, the indiscriminate utilization of all information can easily introduce some irrelevant contents, inevitably hampering performance. In response to this challenge, sparsing, a common information filtering strategy, has been applied in many related studies. Regrettably, their filtering processes often lack reliability and adaptability. To address this issue, we first define an adaptive-reciprocal nearest neighbors (A-RNN) relationship. In identifying neighbors, it gains flexibility through learning adaptive thresholds. In addition, by introducing a reciprocity mechanism, the reliability of neighbors is ensured. Then, we use A-RNN to rectify the similarity matrix in the conventional attention module. In the specific implementation, to distinctly consider non-local and local information, we introduce two blocks: the non-local sparse constraint block and the local sparse constraint block. The former utilizes A-RNN to sparsify non-local information, whereas the latter uses adaptive thresholds to sparsify local information. As a result, an adaptive sparse attention (ASA) module is achieved, inheriting the advantages of flexibility and reliability from A-RNN. In the validation for the proposed ASA module, we use it to replace the attention module in NLNet and conduct experiments on semantic segmentation benchmarks including Cityscapes, ADE20K and PASCAL VOC 2012. With the same backbone (ResNet101), our ASA module outperforms the conventional attention module and its some state-of-the-art variants.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"169 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141509996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Super-resolution reconstruction of images based on residual dual-path interactive fusion combined with attention","authors":"Wang Hao, Peng Taile, Zhou Ying","doi":"10.1117/1.jei.33.3.033034","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.033034","url":null,"abstract":"In recent years, deep learning has made significant progress in the field of single-image super-resolution (SISR) reconstruction, which has greatly improved reconstruction quality. However, most of the SISR networks focus too much on increasing the depth of the network in the process of feature extraction and neglect the connections between different levels of features as well as the full use of low-frequency feature information. To address this problem, this work proposes a network based on residual dual-path interactive fusion combined with attention (RDIFCA). Using the dual interactive fusion strategy, the network achieves the effective fusion and multiplexing of high- and low-frequency information while increasing the depth of the network, which significantly enhances the expressive ability of the network. The experimental results show that the proposed RDIFCA network exhibits certain superiority in terms of objective evaluation indexes and visual effects on the Set5, Set14, BSD100, Urban100, and Manga109 test sets.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"151 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ichraq Lemghari, Sylvie Le Hégarat-Mascle, Emanuel Aldea, Jennifer Vandoni
{"title":"Robust classification with noisy labels using Venn–Abers predictors","authors":"Ichraq Lemghari, Sylvie Le Hégarat-Mascle, Emanuel Aldea, Jennifer Vandoni","doi":"10.1117/1.jei.33.3.031210","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.031210","url":null,"abstract":"The advent of deep learning methods has led to impressive advances in computer vision tasks over the past decades, largely due to their ability to extract non-linear features that are well adapted to the task at hand. For supervised approaches, data labeling is essential to achieve a high level of performance; however, this task can be so fastidious or even troublesome in difficult contexts (e.g., specific defect detection, unconventional data annotations, etc.) that experts can sometimes erroneously provide the wrong ground truth label. Considering classification problems, this paper addresses the issue of handling noisy labels in datasets. Specifically, we first detect the noisy samples of a dataset using set-valued labels and then improve their classification using Venn–Abers predictors. The obtained results reach more than 0.99 and 0.90 accuracy for noisified versions of two widely used image classification datasets, digit MNIST and CIFAR-10 respectively with a 40% two-class pair-flip noise ratio and 0.87 accuracy for CIFAR-10 with 10-class uniform 40% noise ratio.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"12 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Saïd Rahmani, Roger de Souza Lima, Eric Serris, Ana Cameirão, Johan Debayle
{"title":"Monitoring of industrial crystallization processes through image sequence segmentation and characterization","authors":"Saïd Rahmani, Roger de Souza Lima, Eric Serris, Ana Cameirão, Johan Debayle","doi":"10.1117/1.jei.33.3.031211","DOIUrl":"https://doi.org/10.1117/1.jei.33.3.031211","url":null,"abstract":"To enhance control and monitoring of industrial crystallization processes, we propose an innovative nondestructive imaging method utilizing in situ 2D vision sensors. This approach enables the acquisition of 2D videos depicting crystal aggregates throughout the batch crystallization process. Our approach is built upon experimental observations, specifically regarding the process dynamics and sensor fouling. It involves dynamic segmentation of observed aggregates, from which quantitative analyses are derived. Notably, our method allows for tracking the evolution of the particle size distribution of crystal aggregates over time and the determination of the growth kinetics of crystals that agglomerate at the sensor air gap. This enables the detection of key stages in the crystallization process and the geometric characterization of crystal aggregate production.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":"197 1","pages":""},"PeriodicalIF":1.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141518514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}