{"title":"Generic method for grid line detection and removal in scanned documents","authors":"Romain Karpinski, A. Belaïd","doi":"10.1109/ASAR.2018.8480217","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480217","url":null,"abstract":"The detection and extraction of writing grid lines (WGL) in document images is an important task for a wide variety of systems. It is a pre-processing operation that tries to clean up the document image to make the recognition process easier. A lot of work has been proposed for staff line extraction in the context of Optical Music Recognition. Two competitions have been recently proposed in the 2011 and the 2013 ICDAR/GREC conferences. The method proposed in this paper aims to remove WGL without degrading the content. The whole method is based on the estimation of line_space (inter) and line_height and the use of run-length segments to locate WGL points. These points are then grouped together to form larger lines. Missing points are estimated by using a linear model and the context of other adjacent lines. We show that our method does not rely on the writing nature: printed or handwritten nor the language: musical symbols, Latin or Arabic writings. The results obtained are close to the state-of-the-art on not deformed documents. Furthermore, our method performs better than the ones that we have tested (at our disposal) on our image grid datasets.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127332151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Atena Shahkolaei, Azeddine Beghdadi, S. Al-Maadeed, M. Cheriet
{"title":"MHDID: A Multi-distortion Historical Document Image Database","authors":"Atena Shahkolaei, Azeddine Beghdadi, S. Al-Maadeed, M. Cheriet","doi":"10.1109/ASAR.2018.8480372","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480372","url":null,"abstract":"In this paper, a new dataset, called Multi-distortion Historical Document Image Database (MHDID), to be used for the research on quality assessment of degraded documents and degradation classification is proposed. The MHDID dataset contains 335 historical document images which are classified into four categories based on their distortion types, namely, paper translucency, stain, readers’ annotations and worn holes. A total of 36 subjects participated to judge the quality of ancient document images. Pair comparison rating (PCR) is utilized as a subjective rating method for evaluating the visual quality of degraded document images. For each distortion image a mean opinion score (MOS) value is computed. This dataset could be used for evaluating the image quality assessment (IQA) measures as well as in the design of new metrics.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115388108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Novel Term Weighting Scheme and an Approach for Classification of Agricultural Arabic Text Complaints","authors":"D. S. Guru, Mostafa Ali, M. Suhil","doi":"10.1109/ASAR.2018.8480317","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480317","url":null,"abstract":"In this paper, a machine learning based approach for classification of farmers’ complaints which are in Arabic text into different crops has been proposed. Initially, the complaints are preprocessed using stop word removal, auto correction of words, handling some special cases and stemming to extract only the content terms. Some of the domain specific special cases which may affect the classification performance are handled. A new term weighting scheme called Term Class Weight-Inverse Class Frequency (TCW-ICF) is then used to extract the most discriminating features with respect to each class. The extracted features are then used to represent the preprocessed complaints in the form of feature vectors for training a classifier. Finally, an unlabeled complaint is classified as a member of one of the crop classes by the trained classifier. Nevertheless, a relatively large dataset consisting of more than 5000 complaints of the farmers described in Arabic script from eight different crops has been created. The proposed approach has been experimentally validated by conducting an extensive experimentation on the newly created dataset using KNN classifier. It has been argued that the proposed outperforms the baseline Vector Space Model (VSM). Further, the superiority of the proposed term weighting scheme in selecting the best set of discriminating features has been demonstrated through a comparative analysis against four well-known feature selection techniques. The new term is applied on Arabic script as a case study but it can be applied on any text data from any language.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"197 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116393482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Smartphone Arabic Signboards Images Reading","authors":"S. Snoussi","doi":"10.1109/ASAR.2018.8480171","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480171","url":null,"abstract":"In this paper, we present the integration of preprocessing, segmentation and Arabic words recognition system. The obtained system is adapted to be executed by smartphone as an application to help pilgrims (HAJEEJ) from different nationalities to automatically read Arabic signboard images taken by their mobiles and recognize their location. The proposed system involves three main approaches i) an existing approach based on Mathematical Morphology (MM) preprocessing, ii) an Outer Isothetic Cover (OIC) segmentation approach and ii) a Transparent Neural Network (TNN) recognition approach. Note that the proposed system, is a smart one in the way it provides the adequate rules of the next pilgrimage step according to HAJEEJ current position. Hence for such smart system, it would be more fruitful to be suitable not only for desk/lab top machines but mainly for any mobile devices. The proposed system is applied on real database mobile images of specific HAJJ places to evaluate recognition rate, time and memory consuming which are necessary for mobile applications.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126895916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Al-Maadeed, Syed F. K. Peer, Nandhini Subramanian
{"title":"Data Collection and Image Processing System for Ancient Arabic Manuscripts","authors":"S. Al-Maadeed, Syed F. K. Peer, Nandhini Subramanian","doi":"10.1109/ASAR.2018.8480251","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480251","url":null,"abstract":"This paper presents a general-purpose data collection system that combines a DSLR camera with directional LED lamps in order to capture a large quantity of high-resolution manuscript images in such a way as to maximize the speed of data collection while minimizing time and the need for specialized equipment. By integrating custom image processing software, the captured document images are mapped to lie on a planar surface, thereby enabling the application of more sophisticated computer vision algorithms. For this purpose, we also introduce an optional binarization tool that allows researchers to perform basic image pre-processing to simplify later analysis. The hardware setup and software tools presented in this paper can be combined to yield a simple system capable of producing large image datasets for use in document analysis research projects.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134114379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Alaa Abdalhaleem, Berat Kurar Barakat, Jihad El-Sana
{"title":"Case Study: Fine Writing Style Classification Using Siamese Neural Network","authors":"Alaa Abdalhaleem, Berat Kurar Barakat, Jihad El-Sana","doi":"10.1109/ASAR.2018.8480212","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480212","url":null,"abstract":"This paper presents an automatic system for dividing a manuscript into similar parts, according to their similarity in writing style. This system is based on Siamese neural network, which consists of two identical sub-networks joined at their outputs. In the training the two sub-networks extract features from two patches, while the joining neuron measures the distance between the two feature vectors. Patches from the same page are considered as identical and patches from different books are considered as different. Based on that, the Siamese network computes the distances between patches of the same book.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"116 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125720482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Daniel Wilson-Nunn, Terry Lyons, A. Papavasiliou, Hao Ni
{"title":"A Path Signature Approach to Online Arabic Handwriting Recognition","authors":"Daniel Wilson-Nunn, Terry Lyons, A. Papavasiliou, Hao Ni","doi":"10.1109/ASAR.2018.8480300","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480300","url":null,"abstract":"The Arabic script is one that has many properties that come together and result in what is commonly cited as one of the most beautiful scripts. Used by over 400 million people worldwide and with a history spanning over 1800 years, the Arabic script remains one of the most important languages in the world. Using tools from the theory of rough paths, combined with state of the art techniques from deep learning, we develop a recognition methodology for Arabic handwriting. Preliminary results using online Arabic handwritten characters show that the methodology developed can result in a significant decrease in error rate.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129696600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Deep FCN for Arabic Scene Text Detection","authors":"I. Beltaief, Mohamed Ben Halima","doi":"10.1109/ASAR.2018.8480394","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480394","url":null,"abstract":"Visual text is considered as one of the major indispensable aspects of communication field used by individuals and broadly applied in our daily Transactions. Thus, detecting and exploiting this textual information is of a big prominence. State of the art methods for detecting text on printed documents has achieved impressing results on both accuracy and precision values thanks to the sophisticated deep earning approaches, while researchers on natural scenes images still on progress due to the various difficulties on distinguishing text candidates from the remaining shapes. wherefore, as a fast and efficient solution, we propose a deep incorporated multilingual scene text detector system to forthwith localize text using an end-to-end trainable single Network. For training and testing stages, we have used the ACTIV [24] dataset.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123071867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Stephen Rawls, Huaigu Cao, Joe Mathai, P. Natarajan
{"title":"How To Efficiently Increase Resolution in Neural OCR Models","authors":"Stephen Rawls, Huaigu Cao, Joe Mathai, P. Natarajan","doi":"10.1109/ASAR.2018.8480182","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480182","url":null,"abstract":"Modern CRNN OCR models require a fixed line height for all images, and it is known that, up to a point, increasing this input resolution improves recognition performance. However, doing so by simply increasing the line height of input images without changing the CRNN architecture has a large cost in memory and computation (they both scale O(n2) w.r.t. the input line height).We introduce a few very small convolutional and max pooling layers to a CRNN model to rapidly downsample high resolution images to a more manageable resolution before passing off to the \"base\" CRNN model. Doing this greatly improves recognition performance with a very modest increase in computation and memory requirements. We show a 33% relative improvement in WER, from 8.8% to 5.9% when increasing the input resolution from 30px line height to 240px line height on Open-HART/MADCAT Arabic handwriting data.This is a new state of the art result on Arabic handwriting, and the large improvement from an already strong baseline shows the impact of this technique.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130807743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"ASAR 2018 Competition Page Layout Analysis Using Fully Convolutional Networks","authors":"Ahmad Droby, Berat Kurar Barakat, Jihad El-Sana","doi":"10.1109/ASAR.2018.8480326","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480326","url":null,"abstract":"This technical report presents a Fully Convolutional Network based method for layout analysis of benchmarking dataset provided by the competition. The document image is segmented into text and non-text zones by dense pixel prediction. Convolutional part of the network can learn useful features from the document images and is robust to uncontrained layouts. We have evaluated the zone segmentation with average black pixel rate, over-segmentation error, under-segmentation error, correct-segmentation, missed-segmentation error, false alarm error, overall block error rate whereas the zone classification with precision, recall, F1-measure and average class accuracy on both pixel and block levels.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125076430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}