International Conference on Digital Image Processing最新文献

筛选
英文 中文
Feature attention network (FA-Net): a deep-learning based approach for underwater single image enhancement 特征注意网络(FA-Net):一种基于深度学习的水下单幅图像增强方法
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2644516
M. Hamza, Ammar Hawbani, Sami Ul Rehmana, Xingfu Wang, Liang Zhao
{"title":"Feature attention network (FA-Net): a deep-learning based approach for underwater single image enhancement","authors":"M. Hamza, Ammar Hawbani, Sami Ul Rehmana, Xingfu Wang, Liang Zhao","doi":"10.1117/12.2644516","DOIUrl":"https://doi.org/10.1117/12.2644516","url":null,"abstract":"Underwater image processing and analysis have been a hotspot of study in recent years, as more emphasis has been focused to underwater monitoring and usage of marine resources. Compared with the open environment, underwater image encountered with more complicated conditions such as light abortion, scattering, turbulence, nonuniform illumination and color diffusion. Although considerable advances and enhancement techniques achieved in resolving these issues, they treat low-frequency information equally across the entire channel, which results in limiting the network's representativeness. We propose a deep learning and feature-attention-based end-to-end network (FA-Net) to solve this problem. In particular, we propose a Residual Feature Attention Block (RFAB), containing the channel attention, pixel attention, and residual learning mechanism with long and short skip connections. RFAB allows the network to focus on learning high-frequency information while skipping low-frequency information on multi-hop connections. The channel and pixel attention mechanism considers each channel's different features and the uneven distribution of haze over different pixels in the image. The experimental results shows that the FA-Net propose by us provides higher accuracy, quantitatively and qualitatively and superiority to previous state-of-the-art methods.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133625964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
DIRA: disjoint-identity resolution adaptation for low-resolution face recognition 低分辨率人脸识别的分离身份分辨率适应
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2644258
Jacky Chen Long Chai, C. Low, A. Teoh
{"title":"DIRA: disjoint-identity resolution adaptation for low-resolution face recognition","authors":"Jacky Chen Long Chai, C. Low, A. Teoh","doi":"10.1117/12.2644258","DOIUrl":"https://doi.org/10.1117/12.2644258","url":null,"abstract":"Low-resolution face recognition (LRFR) intends to identify unknown poor-quality face images and is widely employed in real-world surveillance applications. While collecting a large-scale labeled low-resolution (LR) face dataset could be conducive, it is practically infeasible due to labor costs and privacy issues. In contrast, accessing high-resolution (HR) face datasets is relatively effortless. However, prevailing domain adaptation techniques are often tenuous as they demand sharing of similar face images at different resolutions. We propose disjoint-identity resolution adaptation (DIRA) to transfer substantial face semantic representations from HR to LR face images, despite disjoint identities and limited labeled LR images. We accredit that continuous adversarial learning between HR-LR resolution alignment and segregation renders effective feature extraction and discriminative LR face representation. Our experimental results show a notable performance boost over the recent state-of-the-art methods for the challenging realistic low-resolution face recognition task.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133897558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detecting roads from high-resolution aerial images: a position iteration algorithm for linear target detection 高分辨率航拍图像道路检测:一种线性目标检测的位置迭代算法
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2644721
Hao He, Shuyang Wang, Qi Yang, Xu Huang, Qian Zhao
{"title":"Detecting roads from high-resolution aerial images: a position iteration algorithm for linear target detection","authors":"Hao He, Shuyang Wang, Qi Yang, Xu Huang, Qian Zhao","doi":"10.1117/12.2644721","DOIUrl":"https://doi.org/10.1117/12.2644721","url":null,"abstract":"Detecting roads from high-resolution photographs can serve forestry, agriculture, traffic and even military areas, and produce significant social and economic value. In this paper, we present a novel method that utilizes the flatness and the connectivity to detect the road in high-resolution aerial images. The method iterates the probable locations of the roads by using the flatness and connects the roads by using the connectivity. Firstly, we introduce a concept of ‘footprint’, which reveals the probable location and extension direction of a road. Given an initial footprint, we assess the flatness between locations to search the resulting footprint. By iterating and connecting the footprints, our approach produces a set of connected line segments that reflect the road to be detected. In addition, a footprints initialization algorithm is introduced to make our method totally automatic, and a road network pruning algorithm is designed to make the result clearer and more accurate. Tested under three high-resolution aerial photographs, our method achieved an accuracy of more than 80%. The algorithm is adapted for road detection and still linear target detection in high-resolution aerial photographs. Since the algorithm does not require artificial features or training data, it can be quickly deployed in application.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134066161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Grad-CAM based visualization of 3D CNNs in classifying fMRI 基于Grad-CAM的三维cnn在fMRI分类中的可视化
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2643867
Jiajun Fu, Meili Lu, Yifan Cao, Zhaohua Guo, Zicheng Gao
{"title":"Grad-CAM based visualization of 3D CNNs in classifying fMRI","authors":"Jiajun Fu, Meili Lu, Yifan Cao, Zhaohua Guo, Zicheng Gao","doi":"10.1117/12.2643867","DOIUrl":"https://doi.org/10.1117/12.2643867","url":null,"abstract":"Deep learning methods have proven promising performance in decoding specific task states based on functional magnetic resonance imaging (fMRI) of the human brain, however, they lack transparency in their decision making, in the sense that it is not straightforward to visualize the features on which the decision was made. In this study, we investigated the decoding of four sensorimotor tasks based on 3D fMRI according to 3D Convolutional Neural Network (3DCNN), and then adopted Grad-CAM algorithms to provide visual explanation from deep networks so as to support the decoding decision.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129662432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparison of different classification methods for autism spectrum diagnosis 自闭症谱系诊断中不同分类方法的比较
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2644418
Shumin Liu, Zhaohui Wang, Linmao Tian, Y. Zhan
{"title":"Comparison of different classification methods for autism spectrum diagnosis","authors":"Shumin Liu, Zhaohui Wang, Linmao Tian, Y. Zhan","doi":"10.1117/12.2644418","DOIUrl":"https://doi.org/10.1117/12.2644418","url":null,"abstract":"Studies have found autism spectrum disorder is a diffuse developmental disease of the central nervous system. The majority of autism cases result from a combination of genetic predisposition and environmental factors that influence early brain development, despite a few being caused by genes alone. Traditional diagnosis of autism spectrum disorder is usually through interviews and questionnaires, which takes plenty of time and might be misdiagnosed. The primary purpose of this study is to compare different classification methods for distinguishing autism spectrum disorder from typical development by machine learning and deep learning in recent years. The experiments are conducted to discuss their strengths and weaknesses, which, in turn, results are presented for further research.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129798904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A NSST-based infrared and visible image fusion method focusing on luminance effect 一种基于nst的聚焦亮度效应的红外与可见光图像融合方法
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2644228
Meng Cai, Xinlong Liu
{"title":"A NSST-based infrared and visible image fusion method focusing on luminance effect","authors":"Meng Cai, Xinlong Liu","doi":"10.1117/12.2644228","DOIUrl":"https://doi.org/10.1117/12.2644228","url":null,"abstract":"Generally, the fused image shows fully the actual situation of the scene and contains more detailed information. However, most fusion methods miss the details of the fusion image and confuses the contrast of the raw scene. To solve this problem, we propose a fusion algorithm based on non-subsampled shearlet transform (NSST) that particularly pays attention to the influence of light intensity when calculating the fusion coefficient. The method first decomposes the input images into high- and low-frequency coefficients through NSST. Then regarding the high-frequency coefficients, we calculate the phase consistency (PC) of the decomposed images, and the results are combined with the adaptive simplified pulse coupled neural network (SPCNN) to compose parameter. Meanwhile, for the low-frequency coefficient, the optimal brightness entropy (OBE) of the input images is obtained as the fusion basis. The next step is to fuse the high- and low-frequency sub-band coefficients by the designed fusion rule, and obtain final image through NSST inverse transformation. Experiments show that our method not only keeps well the image details and maintains the overall image luminance while taking care of the overall effect of the image, but also gets a leading position in some evaluation indicators.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130278353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Surface defect sample generation method based on GAN 基于GAN的表面缺陷样本生成方法
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2644555
Fangyi Ni, Xiaojun Wu, Jinghui Zhou, Zhichang Liu
{"title":"Surface defect sample generation method based on GAN","authors":"Fangyi Ni, Xiaojun Wu, Jinghui Zhou, Zhichang Liu","doi":"10.1117/12.2644555","DOIUrl":"https://doi.org/10.1117/12.2644555","url":null,"abstract":"In order to solve the insufficiency of training data when deep learning technology is applied to surface defect detection task, a surface defect generation algorithm based on generative adversarial network (GAN) was proposed to enhance training sample data. First, a U-shaped convolutional network was designed, and a spatial adaptive normalized structure was introduced to control the mask image to generate the defect shape, and the network from defect-free image to defect image was completed. Second, a multi-layer convolutional discriminant network is designed to extract adversarial feature of the real samples and generated samples. Finally, the adversarial training loss was designed and the generative network adversarial training was completed. Through quantitative contrast experiment, it is proved that the segmentation network has better segmentation results than without data augmentation after using the surface defect generation algorithm to generate data for data augmentation.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132675372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-task model for human pose estimation and person detection 基于多任务模型的人体姿态估计与检测
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2644479
Daiwei Yu, Jun Zhang, Zhao Jin, Guanqun Li, Wenjin Zhang
{"title":"Multi-task model for human pose estimation and person detection","authors":"Daiwei Yu, Jun Zhang, Zhao Jin, Guanqun Li, Wenjin Zhang","doi":"10.1117/12.2644479","DOIUrl":"https://doi.org/10.1117/12.2644479","url":null,"abstract":"Human pose estimation and person detection are two fundamental tasks of human behavior analysis. There has been remarkable progress in these two tasks separately since the development of convolutional neural network. Recently, researchers have paid more attention to one-stage human pose estimation and person detection for the needs of practical application. However, few researches have been reported on completing these two tasks in a single network simultaneously. There are two main reasons: (1) designing an effective mechanism that makes full use of their relevance and complementation to achieve common progress, especially the pose estimation accuracy is really challenging, (2) evaluation bias caused by scale sensitivity difference remains unsolved. To address these problems, we propose a multi-task model for human pose estimation and person detection simultaneously, named PersonPD (person pose and person detection). It predicts keypoint heatmaps and regresses a 4D relative displacement vector (l,t,r,b) which actually encodes the person bounding box and also acts as keypoints' grouping clues. A maximum IOU matching algorithm, named IOU-grouping, is presented to group body joints into individual persons. At the same time, it generates accurate person detection results. In this simple but effective method, our model get competitive person detection and pose estimation performance on COCO datasets1.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126144736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Compressing CNN by alternating constraint optimization framework 交替约束优化框架压缩CNN
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2643734
Peidong Liu, Weirong Liu, Changhong Shi, Zhiqiang Zhang, Zhijun Li, Jie Liu
{"title":"Compressing CNN by alternating constraint optimization framework","authors":"Peidong Liu, Weirong Liu, Changhong Shi, Zhiqiang Zhang, Zhijun Li, Jie Liu","doi":"10.1117/12.2643734","DOIUrl":"https://doi.org/10.1117/12.2643734","url":null,"abstract":"Tensor decomposition has been extensively studied for convolutional neural networks (CNN) model compression. However, the direct decomposition of an uncompressed model into low-rank form causes unavoidable approximation error due to the lack of low-rank property of a pre-trained model. In this manuscript, a CNN model compression method using alternating constraint optimization framework (ACOF) is proposed. Firstly, ACOF formulates tensor decomposition-based model compression as a constraint optimization problem with low tensor rank constraints. This optimization problem is then solved systematically in an iterative manner using alternating direction method of multipliers (ADMM). During the alternating process, the uncompressed model gradually exhibits low-rank tensor property, and then the approximation error in low-rank tensor decomposition can be negligible. Finally, a high-performance CNN compression network can be effectively obtained by SGD-based fine-tuning. Extensive experimental results on image classification show that ACOF produces the optimal compressed model with high performance and low computational complexity. Notably, ACOF compresses Resnet56 to 28% without accuracy drop, and the compressed model have 1.14% higher accuracy than learning-compression (LC) method.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115606069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparative analysis of feature-based methods and direct methods for unmanned system’s vision navigation 基于特征的无人系统视觉导航方法与直接导航方法的比较分析
International Conference on Digital Image Processing Pub Date : 2022-10-12 DOI: 10.1117/12.2643020
Chuanqi Cheng, Haiyan Wang, Jie Huang, Jiaxin Lu, Rui Si
{"title":"Comparative analysis of feature-based methods and direct methods for unmanned system’s vision navigation","authors":"Chuanqi Cheng, Haiyan Wang, Jie Huang, Jiaxin Lu, Rui Si","doi":"10.1117/12.2643020","DOIUrl":"https://doi.org/10.1117/12.2643020","url":null,"abstract":"Vision navigation is an alternative to Global Position System (GPS) in environments where access to GPS is denied, and cameras’ pose estimation is the key technology. At present, the pose estimation methods can be divided into two main techniques: feature-based methods and direct methods. In this paper, we theoretically analyzed the basic principles of feature-based methods and direct methods. The Jacobian matrix of cost function with respect to the pose represented by Lie algebra is derived in detail. Then the nonlinear optimization method is utilized to obtain the optimal camera pose. Finally, the accuracy, real-time and robustness of the two methods are compared and analyzed through systematic and comprehensive experiments.","PeriodicalId":314555,"journal":{"name":"International Conference on Digital Image Processing","volume":"12342 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130956826","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信