{"title":"Nonlinear analysis of video images using deep recurrent auto-associative neural networks for facial understanding","authors":"S. M. Moghadam, S. Seyyedsalehi","doi":"10.1109/PRIA.2017.7983050","DOIUrl":"https://doi.org/10.1109/PRIA.2017.7983050","url":null,"abstract":"Preliminary experiments on the deep architectures of the Auto-Associative Neural Networks demonstrated that they have a fascinating ability in complex nonlinear feature extraction, manifold formation and dimension reduction. However, they should successfully pass a serious challenge of training. Furthermore, using the valuable information inclined in video sequences is so helpful in manifold formation and recognition tasks. Considering sequential information, the recurrent networks are widely used in dynamical modeling. This paper presents a novel nine-layer deep recurrent auto-associative neural network which is capable of simultaneously extracting three different information (identity, emotion and gender) from videos of the face. The proposed framework is extensively evaluated on extended Cohn-Kanade database in analyzing dynamical facial expression. The experimental results demonstrate that the recognition rates of emotion and gender are 95.35% and 97.42%, respectively which is comparable with other state-of-the-art.","PeriodicalId":336066,"journal":{"name":"2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"229 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115691144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A multi-objective component-substitution-based pansharpening","authors":"Ghassem Khademi, H. Ghassemian","doi":"10.1109/PRIA.2017.7983056","DOIUrl":"https://doi.org/10.1109/PRIA.2017.7983056","url":null,"abstract":"This paper proposes a multi-objective approach to improving the component substitution (CS) based pansharpening method by obtaining the adaptive weights. The non-dominated sorting genetic algorithm II (NSGA-II) is employed to simultaneously optimize two objective functions. The inverse of the Correlation Coefficient (CC) and a weighted sum of the Erreur Relative Globale Adimensionnelle de Synthese (ERGAS) in the spectral and spatial domains are used as the objective functions. The use of a multi-objective approach in the CS technique allows optimizing the fused image in terms of both spatial and spectral resolutions. Simulation results show that the proposed method outperforms popular CS-based fusion methods.","PeriodicalId":336066,"journal":{"name":"2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126414722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A new active contour model for tumor segmentation","authors":"Maryam Taghizadeh Dehkordi","doi":"10.1109/PRIA.2017.7983053","DOIUrl":"https://doi.org/10.1109/PRIA.2017.7983053","url":null,"abstract":"In this paper, a new energy function has been proposed for tumor segmentation implemented by the level set method. Multi-scale Gaussian filter is applied to the image and its output determines the probability of each pixel belonging to the tumor structure. Introducing the output into the energy function makes the model robust against the inhomogeneous background. Experimental results from MRI verify the desirable performance of the proposed model in comparison with other methods.","PeriodicalId":336066,"journal":{"name":"2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"115 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117264309","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Vahid Moshkelgosha, Hamed Behzadi-Khormouji, M. Yazdian-Dehkordi
{"title":"Coarse-to-fine parameter tuning for content-based object categorization","authors":"Vahid Moshkelgosha, Hamed Behzadi-Khormouji, M. Yazdian-Dehkordi","doi":"10.1109/PRIA.2017.7983038","DOIUrl":"https://doi.org/10.1109/PRIA.2017.7983038","url":null,"abstract":"Object categorization is an interesting application in computer vision. To develop an efficient system for this purpose, finding an appropriate classifier in conjunction with a suitable feature is essential. Most classifiers and features have one or more parameters to be tuned through cross validation. In this paper, we examined a number of classifiers with several feature descriptors and advise an efficient hybrid feature descriptor for object categorization. Besides, we propose a coarse-to-fine parameter tuning method to avoid exhaustive search within various hyper-parameter of the classifiers. The experimental results provided on a subset of COREL dataset shows the efficiency of the advised hybrid feature and the proposed tuning parameters.","PeriodicalId":336066,"journal":{"name":"2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126706882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Afsoon Asghari Shirazi, A. Dehghani, H. Farsi, M. Yazdi
{"title":"Persian logo recognition using local binary patterns","authors":"Afsoon Asghari Shirazi, A. Dehghani, H. Farsi, M. Yazdi","doi":"10.1109/PRIA.2017.7983058","DOIUrl":"https://doi.org/10.1109/PRIA.2017.7983058","url":null,"abstract":"Nowadays, image processing is getting more popular due to the daily increase of diverse data acquisition methods such as digital scanners and cameras. Due to the high volume of archived documents, automatic document classification methods can help to save the time and space in digital document organization. Logos in official and business documents are used to identify document identities. Different approaches have been used for logo recognition yet, many of which has complex computations to achieve a high level of precision. In this paper, a novel algorithm for accurate logo recognition with low level of computational complexity is proposed based on Local Binary Pattern (LBP). We proposed PerLogo dataset consisting 850 images of 10 different classes of logos has been proposed in this paper. Through 3 separate experiments over 50, 60, 70 images per each class the proposed system has been evaluated. Experimental results show that recognition rate is increased with increasing the number of training images per class. Experimental results show the recognition accuracy of 98% when 0.09 salt and pepper noise are added to the test images, which is more than 95% accuracy proposed by the state-of-the-art approaches achieving 95% accuracy.","PeriodicalId":336066,"journal":{"name":"2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127760425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Classification of power quality events using deep learning on event images","authors":"Ebrahim Balouji, O. Salor","doi":"10.1109/PRIA.2017.7983049","DOIUrl":"https://doi.org/10.1109/PRIA.2017.7983049","url":null,"abstract":"In this paper, a new method for the classification of power quality (PQ) events of the electricity networks based on deep learning approach is presented. In contrast with the existing PQ event data analysis techniques, sampled voltage data of the PQ events are not used, but image files of the three-phase PQ event data are analyzed by taking the advantage of the success of the deep leaning approach on image-file-classification. Therefore, the novelty of the proposed approach is that, image files of the voltage waveforms of the three phases of the power grid are classified. PQ events obtained from four transformer substations of the electricity transmission system for a year are used for training and testing the proposed classification method. DIGITS deep learning platform of NVIDIA is employed for the application of the deep learning algorithm on PQ event data images. It is shown that the test data can be classified with 100% accuracy. The proposed work is believed to serve the needs of the future smart grid applications, which are fast and automatic analysis of the electricity grid and taking automatic countermeasures against potential PQ events.","PeriodicalId":336066,"journal":{"name":"2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133685559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}