2022 IEEE International Conference on Image Processing (ICIP)最新文献

筛选
英文 中文
Monitoring of Varroa Infestation Rate in Beehives: A Simple AI Approach 蜂箱中Varroa侵害率的监测:一种简单的AI方法
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9897809
Lukáš Picek, Adam Novozámský, R. Frydrychová, B. Zitová, Pavel Mach
{"title":"Monitoring of Varroa Infestation Rate in Beehives: A Simple AI Approach","authors":"Lukáš Picek, Adam Novozámský, R. Frydrychová, B. Zitová, Pavel Mach","doi":"10.1109/ICIP46576.2022.9897809","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9897809","url":null,"abstract":"This paper addresses the monitoring of Varroa destructor infestation in Western honey bee colonies. We propose a simple approach using automatic image-based analysis of the fallout on beehive bottom boards. In contrast to the existing high-tech methods, our solution does not require extensive and expensive hardware components, just a standard smart-phone. The described method has the potential to replace the time-consuming, inaccurate, and most common practice where the infestation level is evaluated manually. The underlining machine learning method combines a thresholding algorithm with a shallow CNN—VarroaNet. It provides a reliable estimate of the infestation level with a mean infestation level accuracy of 96.0% and 93.8% in the autumn and winter, respectively. Furthermore, we introduce the developed end-to-end system and its deployment into the online beekeeper’s diary—ProBee—that allows users to identify and track infestation levels on bee colonies.","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":"150 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125959093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-Step Test-Time Adaptation with Entropy Minimization and Pseudo-Labeling 基于熵最小化和伪标记的多步测试时间自适应
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9897419
Hiroaki Kingetsu, Kenichi Kobayashi, Y. Okawa, Yasuto Yokota, K. Nakazawa
{"title":"Multi-Step Test-Time Adaptation with Entropy Minimization and Pseudo-Labeling","authors":"Hiroaki Kingetsu, Kenichi Kobayashi, Y. Okawa, Yasuto Yokota, K. Nakazawa","doi":"10.1109/ICIP46576.2022.9897419","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9897419","url":null,"abstract":"The accuracy of deep neural networks is easily degraded by image corruption. Therefore, there is a need to develop adaptation techniques to ensure durable models and predictions against changes in data distribution. We focus on the task to fit a trained model with a different distribution from training data under the condition that the training data are not available for test time. In this paper, we propose a novel adaptation method in test time for online learning named multi-step layer adaptation (MuSLA). The proposed method achieves high adaptive accuracy by sequentially applying loss functions to specific layers only, especially considering the roles and inter-actions of the layers and employing domain adaptation and semi-supervised learning techniques. The proposed method can be widely applied to already existing trained models with-out additional networks. We show that our approach outperforms conventional methods in image corruption benchmark data experiments.","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121593127","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
The Lifecycle of a Neural Network in the Wild: A Multiple Instance Learning Study on Cancer Detection from Breast Biopsies Imaged with Novel Technique 野外神经网络的生命周期:用新技术从乳腺活检图像中检测癌症的多实例学习研究
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9897596
D. Mandache, E. B. Á. L. Guillaume, Y. Badachi, J. Olivo-Marin, V. Meas-Yedid
{"title":"The Lifecycle of a Neural Network in the Wild: A Multiple Instance Learning Study on Cancer Detection from Breast Biopsies Imaged with Novel Technique","authors":"D. Mandache, E. B. Á. L. Guillaume, Y. Badachi, J. Olivo-Marin, V. Meas-Yedid","doi":"10.1109/ICIP46576.2022.9897596","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9897596","url":null,"abstract":"In the context of tissue examination for breast cancer assessment, we propose a label-free imaging based on Optical Coherence Tomography (OCT) signal combined with a multiple instance learning (MIL) model to respond to a critical need for fast at point-of-care diagnosis: biopsy or surgery time. This new imaging, Dynamic Cell Imaging (DCI), is the time-resolved variant of Full-Field OCT (FFOCT) and offers an intra-cellular resolution of about 1 micron, together with optical sectioning and an improved cell contrast. In order to tackle the challenges of limited data and annotations, while remaining in the scope of interpretability, we design an instance-level MIL model with a focus on adapted data sampling. The interest of this method is that it incorporates task-specific feature learning and also produces instance predictions. For a dataset of 150 core-needle biopsies, we achieve a considerable improvement of more than 20 percentage points in specificity and about 10 in accuracy by leveraging intra-domain (as compared to extra-domain) pre-training.","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115810772","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Background-Tolerant Object Classification With Embedded Segmentation Mask For Infrared And Color Imagery 基于嵌入式分割掩模的红外和彩色图像背景容忍目标分类
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9897418
Maliha Arif, Calvin Yong, Abhijit Mahalanobis, N. Rahnavard
{"title":"Background-Tolerant Object Classification With Embedded Segmentation Mask For Infrared And Color Imagery","authors":"Maliha Arif, Calvin Yong, Abhijit Mahalanobis, N. Rahnavard","doi":"10.1109/ICIP46576.2022.9897418","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9897418","url":null,"abstract":"Even though convolutional neural networks (CNNs) can classify objects in images very accurately, it is well known that the attention of the network may not always be on the semantically important regions of the scene. It has been observed that networks often learn background textures, which are not relevant to the object of interest. In turn this makes the networks susceptible to variations and changes in the background which may negatively affect their performance.We propose a new three-step training procedure called split training to reduce this bias in CNNs for object recognition using Infrared imagery and Color (RGB) data. Our split training procedure has three steps. First, a baseline model is trained to recognize objects in images without background, and the activations produced by the higher layers are observed. Next, a second network is trained using Mean Square Error (MSE) loss to produce the same activations, but in response to the objects embedded in background. This forces the second network to ignore the background while focusing on the object of interest. Finally, with layers producing the activations frozen, the rest of the second network is trained using cross-entropy loss to classify the objects in images with background. Our training method outperforms the traditional training procedure in both a simple CNN architecture, as well as for deep CNNs like VGG and DenseNet, and learns to mimic human vision which focuses more on shape and structure than background with higher accuracy.","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":" 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132075162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Few-Shot Learning Network for Moving Object Detection Using Exemplar-Based Attention Map 基于样例注意图的运动目标检测少镜头学习网络
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9897894
Islam I. Osman, M. Shehata
{"title":"Few-Shot Learning Network for Moving Object Detection Using Exemplar-Based Attention Map","authors":"Islam I. Osman, M. Shehata","doi":"10.1109/ICIP46576.2022.9897894","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9897894","url":null,"abstract":"Moving object detection is a core task in computer vision. However, existing deep learning-based moving object detection methods require a large number of labeled frames to achieve good generalization and performance. This paper proposes a novel deep learning network called FeSh-Net. This network can learn to extract an exemplar-based attention map using a few labeled frames, which guides the network to know which object is foreground and which is a background in the current frame. FeSh-Net is trained using a novel meta-learning technique to be able to segment moving objects from new unseen videos. The proposed network is evaluated using the benchmark CDNet. The results of the proposed FeSh-Net are compared with current state-of-the-art methods, and the results show that FeSh-Net outperforms the best reported state-of-the-art method by 4.4% on average. Additionally, FeSh-Net performs better than other methods when tested using new unseen videos.","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132510727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Towards Lightweight Neural Network-based Chroma Intra Prediction for Video Coding 基于神经网络的视频编码色度内预测研究
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9897708
Chengyi Zou, Shuai Wan, M. Mrak, M. G. Blanch, Luis Herranz, Tiannan Ji
{"title":"Towards Lightweight Neural Network-based Chroma Intra Prediction for Video Coding","authors":"Chengyi Zou, Shuai Wan, M. Mrak, M. G. Blanch, Luis Herranz, Tiannan Ji","doi":"10.1109/ICIP46576.2022.9897708","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9897708","url":null,"abstract":"In video compression the luma channel can be useful for predicting chroma channels (Cb, Cr), as has been demonstrated with the Cross-Component Linear Model (CCLM) used in Versatile Video Coding (VVC) standard. More recently, it has been shown that neural networks can even better capture the relationship among different channels. In this paper, a new attention-based neural network is proposed for cross-component intra prediction. With the goal to simplify neural network design, the new framework consists of four branches: boundary branch and luma branch for extracting features from reference samples, attention branch for fusing the first two branches, and prediction branch for computing the predicted chroma samples. The proposed scheme is integrated into VVC test model together with one additional binary block-level syntax flag which indicates whether a given block makes use of the proposed method. Experimental results demonstrate 0.31%/2.36%/2.00% BD-rate reductions on Y/Cb/Cr components, respectively, on top of the VVC Test Model (VTM) 7.0 which uses CCLM.","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130092558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ICIP 2022 Challenge: PEDCMI, TOOD Enhanced by Slicing-Aided Fine-Tuning and Inference ICIP 2022挑战:通过切片辅助微调和推理增强的PEDCMI, ood
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9897826
Alžběta Turečková, Tomáš Tureček, Z. Oplatková
{"title":"ICIP 2022 Challenge: PEDCMI, TOOD Enhanced by Slicing-Aided Fine-Tuning and Inference","authors":"Alžběta Turečková, Tomáš Tureček, Z. Oplatková","doi":"10.1109/ICIP46576.2022.9897826","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9897826","url":null,"abstract":"This paper describes the approach for the Parasitic Egg Detection and Classification in Microscopic Images challenge. Our solution relies on a robust deep learning pipeline implementing a five-fold training schema to pursue the challenge goal. The final methodology utilizes the TOOD model, further enhanced by slicing-aided fine-tuning and inference. The slicing helps to overcome the image size invariability of the dataset and allows the model to access all images in high resolution, and consequently helps it learn detailed features needed to distinguish different classes and find a precise object position. Our results demonstrate the importance of proper data analysis and consequent pre and post-processing to improve prediction performance.","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130247042","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Channel-Position Self-Attention with Query Refinement Skeleton Graph Neural Network in Human Pose Estimation 基于查询细化骨架图神经网络的通道位置自关注人体姿态估计
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9897882
Shek Wai Chu, Chaoyi Zhang, Yang Song, Weidong (Tom) Cai
{"title":"Channel-Position Self-Attention with Query Refinement Skeleton Graph Neural Network in Human Pose Estimation","authors":"Shek Wai Chu, Chaoyi Zhang, Yang Song, Weidong (Tom) Cai","doi":"10.1109/ICIP46576.2022.9897882","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9897882","url":null,"abstract":"Human Pose Estimation (HPE) is a long-standing yet challenging task in computer vision. The nature of the problem requires comprehensive global contextual reasoning among joints in different locations. In this work, we explore how to incorporate two popular and effective concepts, self-attention and Graph Neural Network (GNN), to model long-range information in HPE. Three different ways to implement self-attention in 3D feature maps are studied, where the best result is achieved via the channel-position version. Accuracy is further improved by refining the queries via an efficient channel-wise parallel GNN that explicitly models the human joint graphical relationships. We are able to improve prediction accuracy on strong baseline models and achieve state-of-the-art results.","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130434335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detection-Identification Balancing Margin Loss for One-Stage Multi-Object Tracking 单阶段多目标跟踪的检测-识别平衡裕度损失
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9898030
Heansung Lee, Suhwan Cho, Sungjun Jang, Jungho Lee, Sungmin Woo, Sangyoun Lee
{"title":"Detection-Identification Balancing Margin Loss for One-Stage Multi-Object Tracking","authors":"Heansung Lee, Suhwan Cho, Sungjun Jang, Jungho Lee, Sungmin Woo, Sangyoun Lee","doi":"10.1109/ICIP46576.2022.9898030","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9898030","url":null,"abstract":"In recent years, one-stage multi-object tracking (MOT) methods, which jointly learn detection and identification in a single network, have attracted extensive attention, due to their efficiency. However, the negative transfer effects caused by the two conflicting objectives of detection and identification have rarely been explored. In this paper, we propose a Detection-Identification Balancing Margin (DIM) loss for minimizing the adverse effects caused by these two different objectives. The proposed DIM loss consists of Detection Margin (DM) loss and Identification Margin (IM) loss. DM loss forces features that are farther from the center of the foreground features than the defined margin due to identification learning to be converged to ensure accurate detection. IM loss enables the various feature representations that are essential for identification by intentionally spreading features that become overly clustered due to detection learning. The proposed DIM loss demonstrates competitive and balanced performance for MOT by providing a positive transfer for features that had a strong negative impact on detection and identification, respectively. (HOTA 61.5, MOTA 75.3, IDF1 75.6 on MOT16, and real-time rates of 25.9 fps were achieved)","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134377863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Viewport-Oriented Panoramic Image Inpainting 面向viewport的全景图像绘制
2022 IEEE International Conference on Image Processing (ICIP) Pub Date : 2022-10-16 DOI: 10.1109/ICIP46576.2022.9897208
Zhuoyi Shang, Yanwei Liu, Guoyi Li, Yunjian Zhang, Jingbo Miao, Jinxia Liu, Liming Wang
{"title":"Viewport-Oriented Panoramic Image Inpainting","authors":"Zhuoyi Shang, Yanwei Liu, Guoyi Li, Yunjian Zhang, Jingbo Miao, Jinxia Liu, Liming Wang","doi":"10.1109/ICIP46576.2022.9897208","DOIUrl":"https://doi.org/10.1109/ICIP46576.2022.9897208","url":null,"abstract":"Panoramic images are usually viewed through Head Mounted Displays (HMDs), which renders only a narrow field of view from the raw panoramic image. This distinctive viewing feature has largely been ignored when inpainting panoramic images. To address this issue, we propose a viewport-oriented generative adversarial panoramic image inpainting network in this paper. For capturing the distorted features accurately in the generating process of equirectangular projection (ERP) panoramic image, a latitude-adaptive feature fusion module is devised to aggregate the latitude-level features in ERP image and less-distorted patch-level viewport-domain features. Furthermore, a novel cross-domain discriminator is proposed to force the inpainting network to generate more plausible results in viewports. Extensive experiments show that our model achieves better performance compared to the baseline methods, especially in the viewport images.","PeriodicalId":387035,"journal":{"name":"2022 IEEE International Conference on Image Processing (ICIP)","volume":"30 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131746541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信