2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)最新文献

筛选
英文 中文
Generic method for grid line detection and removal in scanned documents 扫描文件中网格线检测和去除的通用方法
Romain Karpinski, A. Belaïd
{"title":"Generic method for grid line detection and removal in scanned documents","authors":"Romain Karpinski, A. Belaïd","doi":"10.1109/ASAR.2018.8480217","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480217","url":null,"abstract":"The detection and extraction of writing grid lines (WGL) in document images is an important task for a wide variety of systems. It is a pre-processing operation that tries to clean up the document image to make the recognition process easier. A lot of work has been proposed for staff line extraction in the context of Optical Music Recognition. Two competitions have been recently proposed in the 2011 and the 2013 ICDAR/GREC conferences. The method proposed in this paper aims to remove WGL without degrading the content. The whole method is based on the estimation of line_space (inter) and line_height and the use of run-length segments to locate WGL points. These points are then grouped together to form larger lines. Missing points are estimated by using a linear model and the context of other adjacent lines. We show that our method does not rely on the writing nature: printed or handwritten nor the language: musical symbols, Latin or Arabic writings. The results obtained are close to the state-of-the-art on not deformed documents. Furthermore, our method performs better than the ones that we have tested (at our disposal) on our image grid datasets.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127332151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MHDID: A Multi-distortion Historical Document Image Database MHDID:一个多失真历史文档图像数据库
Atena Shahkolaei, Azeddine Beghdadi, S. Al-Maadeed, M. Cheriet
{"title":"MHDID: A Multi-distortion Historical Document Image Database","authors":"Atena Shahkolaei, Azeddine Beghdadi, S. Al-Maadeed, M. Cheriet","doi":"10.1109/ASAR.2018.8480372","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480372","url":null,"abstract":"In this paper, a new dataset, called Multi-distortion Historical Document Image Database (MHDID), to be used for the research on quality assessment of degraded documents and degradation classification is proposed. The MHDID dataset contains 335 historical document images which are classified into four categories based on their distortion types, namely, paper translucency, stain, readers’ annotations and worn holes. A total of 36 subjects participated to judge the quality of ancient document images. Pair comparison rating (PCR) is utilized as a subjective rating method for evaluating the visual quality of degraded document images. For each distortion image a mean opinion score (MOS) value is computed. This dataset could be used for evaluating the image quality assessment (IQA) measures as well as in the design of new metrics.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115388108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Novel Term Weighting Scheme and an Approach for Classification of Agricultural Arabic Text Complaints 一种新的阿拉伯语农业文本投诉术语加权方案及分类方法
D. S. Guru, Mostafa Ali, M. Suhil
{"title":"A Novel Term Weighting Scheme and an Approach for Classification of Agricultural Arabic Text Complaints","authors":"D. S. Guru, Mostafa Ali, M. Suhil","doi":"10.1109/ASAR.2018.8480317","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480317","url":null,"abstract":"In this paper, a machine learning based approach for classification of farmers’ complaints which are in Arabic text into different crops has been proposed. Initially, the complaints are preprocessed using stop word removal, auto correction of words, handling some special cases and stemming to extract only the content terms. Some of the domain specific special cases which may affect the classification performance are handled. A new term weighting scheme called Term Class Weight-Inverse Class Frequency (TCW-ICF) is then used to extract the most discriminating features with respect to each class. The extracted features are then used to represent the preprocessed complaints in the form of feature vectors for training a classifier. Finally, an unlabeled complaint is classified as a member of one of the crop classes by the trained classifier. Nevertheless, a relatively large dataset consisting of more than 5000 complaints of the farmers described in Arabic script from eight different crops has been created. The proposed approach has been experimentally validated by conducting an extensive experimentation on the newly created dataset using KNN classifier. It has been argued that the proposed outperforms the baseline Vector Space Model (VSM). Further, the superiority of the proposed term weighting scheme in selecting the best set of discriminating features has been demonstrated through a comparative analysis against four well-known feature selection techniques. The new term is applied on Arabic script as a case study but it can be applied on any text data from any language.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"197 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116393482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Smartphone Arabic Signboards Images Reading 智能手机阿拉伯语招牌图像阅读
S. Snoussi
{"title":"Smartphone Arabic Signboards Images Reading","authors":"S. Snoussi","doi":"10.1109/ASAR.2018.8480171","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480171","url":null,"abstract":"In this paper, we present the integration of preprocessing, segmentation and Arabic words recognition system. The obtained system is adapted to be executed by smartphone as an application to help pilgrims (HAJEEJ) from different nationalities to automatically read Arabic signboard images taken by their mobiles and recognize their location. The proposed system involves three main approaches i) an existing approach based on Mathematical Morphology (MM) preprocessing, ii) an Outer Isothetic Cover (OIC) segmentation approach and ii) a Transparent Neural Network (TNN) recognition approach. Note that the proposed system, is a smart one in the way it provides the adequate rules of the next pilgrimage step according to HAJEEJ current position. Hence for such smart system, it would be more fruitful to be suitable not only for desk/lab top machines but mainly for any mobile devices. The proposed system is applied on real database mobile images of specific HAJJ places to evaluate recognition rate, time and memory consuming which are necessary for mobile applications.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126895916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Data Collection and Image Processing System for Ancient Arabic Manuscripts 古阿拉伯语手稿数据采集与图像处理系统
S. Al-Maadeed, Syed F. K. Peer, Nandhini Subramanian
{"title":"Data Collection and Image Processing System for Ancient Arabic Manuscripts","authors":"S. Al-Maadeed, Syed F. K. Peer, Nandhini Subramanian","doi":"10.1109/ASAR.2018.8480251","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480251","url":null,"abstract":"This paper presents a general-purpose data collection system that combines a DSLR camera with directional LED lamps in order to capture a large quantity of high-resolution manuscript images in such a way as to maximize the speed of data collection while minimizing time and the need for specialized equipment. By integrating custom image processing software, the captured document images are mapped to lie on a planar surface, thereby enabling the application of more sophisticated computer vision algorithms. For this purpose, we also introduce an optional binarization tool that allows researchers to perform basic image pre-processing to simplify later analysis. The hardware setup and software tools presented in this paper can be combined to yield a simple system capable of producing large image datasets for use in document analysis research projects.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134114379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Case Study: Fine Writing Style Classification Using Siamese Neural Network 案例研究:使用暹罗神经网络进行精细写作风格分类
Alaa Abdalhaleem, Berat Kurar Barakat, Jihad El-Sana
{"title":"Case Study: Fine Writing Style Classification Using Siamese Neural Network","authors":"Alaa Abdalhaleem, Berat Kurar Barakat, Jihad El-Sana","doi":"10.1109/ASAR.2018.8480212","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480212","url":null,"abstract":"This paper presents an automatic system for dividing a manuscript into similar parts, according to their similarity in writing style. This system is based on Siamese neural network, which consists of two identical sub-networks joined at their outputs. In the training the two sub-networks extract features from two patches, while the joining neuron measures the distance between the two feature vectors. Patches from the same page are considered as identical and patches from different books are considered as different. Based on that, the Siamese network computes the distances between patches of the same book.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"116 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125720482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A Path Signature Approach to Online Arabic Handwriting Recognition 一种路径签名方法用于在线阿拉伯手写识别
Daniel Wilson-Nunn, Terry Lyons, A. Papavasiliou, Hao Ni
{"title":"A Path Signature Approach to Online Arabic Handwriting Recognition","authors":"Daniel Wilson-Nunn, Terry Lyons, A. Papavasiliou, Hao Ni","doi":"10.1109/ASAR.2018.8480300","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480300","url":null,"abstract":"The Arabic script is one that has many properties that come together and result in what is commonly cited as one of the most beautiful scripts. Used by over 400 million people worldwide and with a history spanning over 1800 years, the Arabic script remains one of the most important languages in the world. Using tools from the theory of rough paths, combined with state of the art techniques from deep learning, we develop a recognition methodology for Arabic handwriting. Preliminary results using online Arabic handwritten characters show that the methodology developed can result in a significant decrease in error rate.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129696600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Deep FCN for Arabic Scene Text Detection 用于阿拉伯语场景文本检测的深度FCN
I. Beltaief, Mohamed Ben Halima
{"title":"Deep FCN for Arabic Scene Text Detection","authors":"I. Beltaief, Mohamed Ben Halima","doi":"10.1109/ASAR.2018.8480394","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480394","url":null,"abstract":"Visual text is considered as one of the major indispensable aspects of communication field used by individuals and broadly applied in our daily Transactions. Thus, detecting and exploiting this textual information is of a big prominence. State of the art methods for detecting text on printed documents has achieved impressing results on both accuracy and precision values thanks to the sophisticated deep earning approaches, while researchers on natural scenes images still on progress due to the various difficulties on distinguishing text candidates from the remaining shapes. wherefore, as a fast and efficient solution, we propose a deep incorporated multilingual scene text detector system to forthwith localize text using an end-to-end trainable single Network. For training and testing stages, we have used the ACTIV [24] dataset.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123071867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
How To Efficiently Increase Resolution in Neural OCR Models 如何有效提高神经OCR模型的分辨率
Stephen Rawls, Huaigu Cao, Joe Mathai, P. Natarajan
{"title":"How To Efficiently Increase Resolution in Neural OCR Models","authors":"Stephen Rawls, Huaigu Cao, Joe Mathai, P. Natarajan","doi":"10.1109/ASAR.2018.8480182","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480182","url":null,"abstract":"Modern CRNN OCR models require a fixed line height for all images, and it is known that, up to a point, increasing this input resolution improves recognition performance. However, doing so by simply increasing the line height of input images without changing the CRNN architecture has a large cost in memory and computation (they both scale O(n2) w.r.t. the input line height).We introduce a few very small convolutional and max pooling layers to a CRNN model to rapidly downsample high resolution images to a more manageable resolution before passing off to the \"base\" CRNN model. Doing this greatly improves recognition performance with a very modest increase in computation and memory requirements. We show a 33% relative improvement in WER, from 8.8% to 5.9% when increasing the input resolution from 30px line height to 240px line height on Open-HART/MADCAT Arabic handwriting data.This is a new state of the art result on Arabic handwriting, and the large improvement from an already strong baseline shows the impact of this technique.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130807743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
ASAR 2018 Competition Page Layout Analysis Using Fully Convolutional Networks 基于全卷积网络的ASAR 2018竞赛页面布局分析
Ahmad Droby, Berat Kurar Barakat, Jihad El-Sana
{"title":"ASAR 2018 Competition Page Layout Analysis Using Fully Convolutional Networks","authors":"Ahmad Droby, Berat Kurar Barakat, Jihad El-Sana","doi":"10.1109/ASAR.2018.8480326","DOIUrl":"https://doi.org/10.1109/ASAR.2018.8480326","url":null,"abstract":"This technical report presents a Fully Convolutional Network based method for layout analysis of benchmarking dataset provided by the competition. The document image is segmented into text and non-text zones by dense pixel prediction. Convolutional part of the network can learn useful features from the document images and is robust to uncontrained layouts. We have evaluated the zone segmentation with average black pixel rate, over-segmentation error, under-segmentation error, correct-segmentation, missed-segmentation error, false alarm error, overall block error rate whereas the zone classification with precision, recall, F1-measure and average class accuracy on both pixel and block levels.","PeriodicalId":165564,"journal":{"name":"2018 IEEE 2nd International Workshop on Arabic and Derived Script Analysis and Recognition (ASAR)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125076430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信