2020 International Conference on Machine Vision and Image Processing (MVIP)最新文献

筛选
英文 中文
Scale Equivariant CNNs with Scale Steerable Filters 具有尺度可控制滤波器的尺度等变cnn
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116889
Hanieh Naderi, Leili Goli, S. Kasaei
{"title":"Scale Equivariant CNNs with Scale Steerable Filters","authors":"Hanieh Naderi, Leili Goli, S. Kasaei","doi":"10.1109/MVIP49855.2020.9116889","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116889","url":null,"abstract":"Convolution Neural Networks (CNNs), despite being one of the most successful image classification methods, are not robust to most geometric transformations (rotation, isotropic scaling) because of their structural constraints. Recently, scale steerable filters have been proposed to allow scale invariance in CNNs. Although these filters enhance the network performance in scaled image classification tasks, they cannot maintain the scale information across the network. In this paper, this problem is addressed. First, a CNN is built with the usage of scale steerable filters. Then, a scale equivariat network is acquired by adding a feature map to each layer so that the scale-related features are retained across the network. At last, by defining the cost function as the cross entropy, this solution is evaluated and the model parameters are updated. The results show that it improves the perfromance about 2% over other comparable methods of scale equivariance and scale invariance, when run on the FMNIST-scale dataset.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126383383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Extracting Iso-Disparity Strip Width using a Statistical Model in a Stereo Vision System 基于统计模型的立体视觉等视差条带宽度提取
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116926
Benyamin Kheradvar, A. Mousavinia, A. M. Sodagar
{"title":"Extracting Iso-Disparity Strip Width using a Statistical Model in a Stereo Vision System","authors":"Benyamin Kheradvar, A. Mousavinia, A. M. Sodagar","doi":"10.1109/MVIP49855.2020.9116926","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116926","url":null,"abstract":"Disparity map images, as outputs of a stereo vision system, are known as an effective approach in applications that need depth information in their procedure. One example of such applications is extracting planes with arbitrary attributes from a scene using the concept of iso-disparity strips. The width and direction of strips depend on the plane direction and position in the 3D space. In this paper, a statistical analysis is performed to model the behavior of these strips. This statistical analysis as well as a frequency analysis reveal that for each group of iso-disparity strips, which are corresponding to a single plane in 3D, the width of strips can be represented by an average value superposed by an Additive Gaussian Noise (AGN). This means that a simple averaging technique can significantly reduce the measurement noise in applications such as ground detection using these strips. Results show that the width of iso-disparity strips can be measured with an average precision of 96% using the presented noise model.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122205070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MVIP 2020 Table of Authors MVIP 2020作者表
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/mvip49855.2020.9116925
{"title":"MVIP 2020 Table of Authors","authors":"","doi":"10.1109/mvip49855.2020.9116925","DOIUrl":"https://doi.org/10.1109/mvip49855.2020.9116925","url":null,"abstract":"","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"117 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132638494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using Siamese Networks with Transfer Learning for Face Recognition on Small-Samples Datasets 基于Siamese网络和迁移学习的小样本人脸识别
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116915
Mohsen Heidari, Kazim Fouladi-Ghaleh
{"title":"Using Siamese Networks with Transfer Learning for Face Recognition on Small-Samples Datasets","authors":"Mohsen Heidari, Kazim Fouladi-Ghaleh","doi":"10.1109/MVIP49855.2020.9116915","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116915","url":null,"abstract":"Nowadays, computer-based face recognition is a mature and reliable mechanism that is significantly used in many access control scenarios along with other biometric methods. Face recognition consists of two subtasks including Face Verification and Face Identification. By comparing a pair of images, Face Verification determines whether those images are related to one person or not; and Face Identification has to identify a specific face within a set of available faces in the database. There are many challenges in face recognition such as angle, illumination, pose, facial expression, noise, resolution, occlusion and the few number of one-class samples with several classes. In this paper, we are carrying out face recognition by utilizing transfer learning in a siamese network which consists of two similar CNNs. In the siamese network, a pair of two face images is given to the network as input, then the network extracts the features of this pair of images and finally, it determines whether the pair of images belongs to one person or not by using a similarity criterion. The results show that the proposed model is comparable with advanced models that are trained on datasets containing large numbers of samples. furthermore, it improves the accuracy of face recognition in comparison with methods which are trained using datasets with a few number of samples, and the mentioned accuracy is claimed to be 95.62% on LFW dataset.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114166227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
DeepFaceAR: Deep Face Recognition and Displaying Personal Information via Augmented Reality DeepFaceAR:通过增强现实技术进行深度人脸识别和显示个人信息
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116873
Amin Golnari, H. Khosravi, S. Sanei
{"title":"DeepFaceAR: Deep Face Recognition and Displaying Personal Information via Augmented Reality","authors":"Amin Golnari, H. Khosravi, S. Sanei","doi":"10.1109/MVIP49855.2020.9116873","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116873","url":null,"abstract":"Biometric recognition is a popular topic in machine vision. Deep Neural Networks have been recently used in several applications, especially in biometric recognition. In this paper, we combine a Deep Neural Network alongside Augmented Reality to produce a system capable of recognizing the faces of individuals and displaying some information about the individual as an Augmented Reality. We used a dataset containing 1200 face images of 100 faculty members of the Shahrood University of Technology. After training the proposed Deep Network, it reached the recognition accuracy of 99.45%. We also provided some graphical targets for each person that contains his information. When a person is identified by the deep network, the target image provided for augmented reality is aligned with the angle and dimensions of the detected face and displayed on top of it.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126131215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Modeling of Pruning Techniques for Simplifying Deep Neural Networks 简化深度神经网络的剪枝技术建模
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116891
Morteza Mousa Pasandi, M. Hajabdollahi, N. Karimi, S. Samavi
{"title":"Modeling of Pruning Techniques for Simplifying Deep Neural Networks","authors":"Morteza Mousa Pasandi, M. Hajabdollahi, N. Karimi, S. Samavi","doi":"10.1109/MVIP49855.2020.9116891","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116891","url":null,"abstract":"Convolutional Neural Networks (CNNs) suffer from different issues such as computational complexity and the number of parameters. In recent years pruning techniques are employed to reduce the number of operations and model size in CNNs. Different pruning methods are proposed, which are based on pruning the connections, channels, and filters. Various techniques and tricks accompany pruning methods, and there is not a unifying framework to model all the pruning methods. In this paper pruning methods are investigated, and a general model which is contained the majority of pruning techniques is proposed. The advantages and disadvantages of the pruning methods can be identified, and all of them can be summarized under this model. The final goal of this model can be providing a specific method for all the pruning methods with different structures and applications.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124848469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Image Watermarking by Q Learning and Matrix Factorization 基于Q学习和矩阵分解的图像水印
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116871
M. Alizadeh, H. Sajedi, B. BabaAli
{"title":"Image Watermarking by Q Learning and Matrix Factorization","authors":"M. Alizadeh, H. Sajedi, B. BabaAli","doi":"10.1109/MVIP49855.2020.9116871","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116871","url":null,"abstract":"Today, with the advancement of technology and the widespread use of the internet, watermarking techniques are being developed to protect copyright and data security. The methods proposed for watermarking can be divided into two main categories: spatial domain watermarking, and frequency domain watermarking. Often matrix transformation methods are merged with another method to select the right place to hide. In this paper, a non-blind watermarking id presented. In order to embed watermark Least Significant Bit (LSB) replacement and QR matrix factorization are exploited. Q learning is used to select the appropriate host blocks. The Peak Signal-to-Noise Ratio(PSNR) of the watermarked image and the extracted watermark image is considered as the reward function. The proposed method has been improved over the algorithms mentioned above with no learning methods and achieved a mean PSNR values of 56.61 dB and 55.77 dB for QR matrix factorization and LSB replacemnet embedding method respectively.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120921028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Occluded Visual Object Recognition Using Deep Conditional Generative Adversarial Nets and Feedforward Convolutional Neural Networks 基于深度条件生成对抗网络和前馈卷积神经网络的遮挡视觉目标识别
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116887
Vahid Reza Khazaie, Alireza Akhavanpour, R. Ebrahimpour
{"title":"Occluded Visual Object Recognition Using Deep Conditional Generative Adversarial Nets and Feedforward Convolutional Neural Networks","authors":"Vahid Reza Khazaie, Alireza Akhavanpour, R. Ebrahimpour","doi":"10.1109/MVIP49855.2020.9116887","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116887","url":null,"abstract":"Core object recognition is the task of recognizing objects without regard to any variations in the conditions like pose, illumination or any other structural modifications. This task is solved through the feedforward processing of information in the human visual system. Deep neural networks can perform like humans in this task. However, we do not know how object recognition under more challenging conditions like occlusion is solved. Some computational models imply that recurrent processing might be a solution to the beyond core object recognition task. The other potential mechanism for solving occlusion is to reconstruct the occluded part of the object taking advantage of generative models. Here we used Conditional Generative Adversarial Networks for reconstruction. For reasonable size occlusion, we were able to remove the effect of occlusion and we recovered the performance of the base model. We showed getting the benefit of GANs for reconstruction and adding information by generative models can cause a better performance in the object recognition task under occlusion.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131118549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Classified and Comparative Study of 2-D Convolvers 二维卷积的分类与比较研究
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116874
Mahdi Kalbasi, Hooman Nikmehr
{"title":"A Classified and Comparative Study of 2-D Convolvers","authors":"Mahdi Kalbasi, Hooman Nikmehr","doi":"10.1109/MVIP49855.2020.9116874","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116874","url":null,"abstract":"Two-dimensional (2-D) convolution is a common operation in a wide range of signal and image processing applications such as edge detection, sharpening, and blurring. In the hardware implementation of these applications, 2d convolution is one of the most challenging parts because it is a compute-intensive and memory-intensive operation. To address these challenges, several design techniques such as pipelining, constant multiplication, and time-sharing have been applied in the literature which leads to convolvers with different implementation features. In this paper, based on design techniques, we classify these convolvers into four classes named Non-Pipelined Convolver, Reduced-Bandwidth Pipelined Convolver, Multiplier-Less Pipelined Convolver, and Time-Shared Convolver. Then, implementation features of these classes, such as critical path delay, memory bandwidth, and resource utilization, are analyticcally discussed for different convolution kernel sizes. Finally, an instance of each class is captured in Verilog and their features are evaluated by implementing them on a Virtex-7 FPGA and reported confirming the analytical discussions.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130265463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
MVIP 2020 Cover Page MVIP 2020封面
2020 International Conference on Machine Vision and Image Processing (MVIP) Pub Date : 2020-02-01 DOI: 10.1109/mvip49855.2020.9116892
{"title":"MVIP 2020 Cover Page","authors":"","doi":"10.1109/mvip49855.2020.9116892","DOIUrl":"https://doi.org/10.1109/mvip49855.2020.9116892","url":null,"abstract":"","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134143798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信