The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501133
George V. Landon
{"title":"Automatic photometric restoration of historical photographic negatives","authors":"George V. Landon","doi":"10.1145/2501115.2501133","DOIUrl":"https://doi.org/10.1145/2501115.2501133","url":null,"abstract":"The majority of early photographs were captured on acetate-based film. However, it has been determined that these negatives will deteriorate beyond repair even with proper conservation and no suitable restoration method is available without physically altering each negative. In this paper, we present an automatic method to remove various nonlinear illumination distortions caused by deteriorating photographic support material. First, using a High-Dynamic Range structured-light scanning method, a 2D Gaussian model for light transmission is estimated for each pixel of the negative image. Estimated amplitude at each pixel provides an accurate model of light transmission, but also includes regions of lower transmission caused by damaged areas. Principal Component Analysis is then used to estimate the photometric error and effectively restore the original illumination information of the negative. Using both the shift in the Gaussian light stripes between pixels and their variations in standard deviation, a 3D surface estimate is calculated. Experiments of real historical negatives show promising results for widespread implementation in memory institutions.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"29 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80030476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501126
William B. Lund, Douglas J. Kennard, Eric K. Ringger
{"title":"Why multiple document image binarizations improve OCR","authors":"William B. Lund, Douglas J. Kennard, Eric K. Ringger","doi":"10.1145/2501115.2501126","DOIUrl":"https://doi.org/10.1145/2501115.2501126","url":null,"abstract":"Our previous work has shown that the error correction of optical character recognition (OCR) on degraded historical machine-printed documents is improved with the use of multiple information sources and multiple OCR hypotheses including from multiple document image binarizations. The contributions of this paper are in demonstrating how diversity among multiple binarizations makes those improvements to OCR accuracy possible. We demonstrate the degree and breadth to which the information required for correction is distributed across multiple binarizations of a given document image. Our analysis reveals that the sources of these corrections are not limited to any single binarization and that the full range of binarizations holds information needed to achieve the best result as measured by the word error rate (WER) of the final OCR decision. Even binarizations with high WERs contribute to improving the final OCR. For the corpus used in this research, fully 2.68% of all tokens are corrected using hypotheses not found in the OCR of the binarized image with the lowest WER. Further, we show that the higher the WER of the OCR overall, the more the corrections are distributed among all binarizations of the document image.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"15 1","pages":"86-93"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84375702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501128
Raid Saabni
{"title":"The multi angular descriptor (MAD): a binary and gray images descriptor for shape recognition","authors":"Raid Saabni","doi":"10.1145/2501115.2501128","DOIUrl":"https://doi.org/10.1145/2501115.2501128","url":null,"abstract":"In this paper, we present the Multi Angular Descriptor (MAD), a new shape descriptor for shape based object recognition and image retrieval. In the binary case, the MAD descriptor captures the angular view to multi resolution rings from each contour point. Placing the rings in different heights enables capturing multi-level global/local features. In gray level, it captures the weighted distribution over relative positions of the shape points to multi resolution rings around the centroid. The multi angular descriptor is robust to noise and small deformations. Flexible parameters makes the MAD descriptor tunable to specific unique characteristics of the different tasks. The extension of the (MAD) descriptor to gray level shapes, can be seen as an extension of a shape context descriptor to be used with low quality gray level images avoiding poor results of the binarization process. Testing the proposed descriptor on the MNIST dataset [16] and a private dataset using two matching techniques gave better results comparing to the Shapes Context and the Histogram of Oriented Gradients (HOG) descriptors.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"92 1","pages":"53-58"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91111327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501127
V. C. Kieu, M. Visani, N. Journet, R. Mullot, J. Domenger
{"title":"An efficient parametrization of character degradation model for semi-synthetic image generation","authors":"V. C. Kieu, M. Visani, N. Journet, R. Mullot, J. Domenger","doi":"10.1145/2501115.2501127","DOIUrl":"https://doi.org/10.1145/2501115.2501127","url":null,"abstract":"This paper presents an efficient parametrization method for generating synthetic noise on document images. By specifying the desired categories and amount of noise, the method is able to generate synthetic document images with most of degradations observed in real document images (ink splotches, white specks or streaks). Thanks to the ability of simulating different amount and kind of noise, it is possible to evaluate the robustness of many document image analysis methods. It also permits to generate data for algorithms that employ a learning process. The degradation model presented in [7] needs eight parameters for generating randomly noise regions. We propose here an extension of this model which aims to set automatically the eight parameters to generate precisely what a user wants (amount and category of noise). Our proposition consists of three steps. First, Nsp seed-points (i.e. centres of noise regions) are selected by an adaptive procedure. Then, these seed-points are classified into three categories of noise by using a heuristic rule. Finally, each size of noise region is set using a random process in order to generate degradations as realistic as possible.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"1 1","pages":"29-35"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89828063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501129
Chulapong Panichkriangkrai, Liang Li, K. Hachimura
{"title":"Character segmentation and retrieval for learning support system of Japanese historical books","authors":"Chulapong Panichkriangkrai, Liang Li, K. Hachimura","doi":"10.1145/2501115.2501129","DOIUrl":"https://doi.org/10.1145/2501115.2501129","url":null,"abstract":"This paper proposes a character segmentation and retrieval method for a learning support system that analyzes digitized Japanese historical woodblock printed books. The proposed system detects text lines, segments characters, and retrieves similar characters from document images. The process includes background separation, text line extraction, rule-based character integration and segmentation, and similar character retrieval. The experimental results show that the proposed method segmented all text lines correctly and successfully extracted more than 79% of the complicated characters, as well as provided promising character retrieval results.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"1 1","pages":"118-122"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91203899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501132
T. Packer, D. Embley
{"title":"Cost effective ontology population with data from lists in OCRed historical documents","authors":"T. Packer, D. Embley","doi":"10.1145/2501115.2501132","DOIUrl":"https://doi.org/10.1145/2501115.2501132","url":null,"abstract":"A method of automatically extracting facts from lists in OCRed documents and inserting them into an ontology would contribute to making a variety of historical knowledge machine searchable, queryable, and linkable. To work well, such a process must be adaptable to variations in list format, tolerant of OCR errors, and careful in its selection of human guidance. We propose ListReader, a wrapper-induction solution for information extraction that is specialized for lists in OCRed documents. ListReader can induce either a regular-expression grammar or a Hidden Markov Model. Each can infer list structure and field labels from OCR text. We decrease the cost and improve the accuracy of the induction process using semi-supervised machine learning and active learning, allowing induction of a wrapper from almost a single hand-labeled instance per field per list. After applying an induced wrapper, ListReader automatically maps the labeled text it produces to a rich variety of ontologically structured predicates. We evaluate our implementation on family history books in terms of the typical F-measure and a new metric, \"Label Efficiency\", which measures both extraction quality and cost in a single number. We show with statistical significance that ListReader reaches values closer to optimal levels than a state-of-the-art statistical sequence labeler.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"1 1","pages":"44-52"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90927628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501122
T. V. Phan, Hajime Baba, Akihiro Watanabe, M. Nakagawa
{"title":"A re-assembling scheme of fragmented Mokkan images","authors":"T. V. Phan, Hajime Baba, Akihiro Watanabe, M. Nakagawa","doi":"10.1145/2501115.2501122","DOIUrl":"https://doi.org/10.1145/2501115.2501122","url":null,"abstract":"Historical documents are invaluable to study the society and culture in old ages everywhere in the world. In Japan, unearthed wooden tablets called Mokkan excavated from ancient palace sites and so on in the Nara period provide important clues to know the era. Since most of unearthed Mokkan have been badly damaged and broken into several pieces, however, it is extremely difficult even for experts to extract characters on fragmented Mokkan. In this paper, we propose a digital image reassembling scheme for fragmented Mokkan so that broken character images are reassembled and written content is analyzed. The proposed scheme consists of two steps: an image grouping using color features and an image reassembling using local tangent and curvature functions of the fragment contours. After the grouping process, fragment images with the same color features are clustered. Then, in the reassembling step, candidate matching pairs for adjacent fragment images in the same group are listed. We also provide a user interface for archeologists to verify the results. As a result, the system helps archaeologists reconstruct Mokkan images so that they can decode them.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"201 1","pages":"22-28"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86835318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501124
Róisín Rowley-Brooke, François Pitié, A. Kokaram
{"title":"Nonrigid recto-verso registration using page outline structure and content preserving warps","authors":"Róisín Rowley-Brooke, François Pitié, A. Kokaram","doi":"10.1145/2501115.2501124","DOIUrl":"https://doi.org/10.1145/2501115.2501124","url":null,"abstract":"Accurate registration of document recto and verso sides with bleed-through degradation is essential for accurate automatic non-blind bleed-through removal. This paper presents a registration method for documents with bleed-through degradation, and also an objective registration evaluation scheme. In the proposed method the two sides are first globally aligned using the outline of the page, and then a local grid point warp is applied using the sum of squared difference between both the image intensity and gradient fields as an error metric, and a content preserving smoothness penalty. The displacement fields of the combined global and local warps are then evaluated against manually registered full document image displacement fields and compared with recent document registration methods.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"3 1","pages":"8-13"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81903431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501121
Maroua Mehri, Petra Gomez-Krämer, P. Héroux, A. Boucher, R. Mullot
{"title":"Texture feature evaluation for segmentation of historical document images","authors":"Maroua Mehri, Petra Gomez-Krämer, P. Héroux, A. Boucher, R. Mullot","doi":"10.1145/2501115.2501121","DOIUrl":"https://doi.org/10.1145/2501115.2501121","url":null,"abstract":"Texture feature analysis has undergone tremendous growth in recent years. It plays an important role for the analysis of many kinds of images. More recently, the use of texture analysis techniques for historical document image segmentation has become a logical and relevant choice in the conditions of significant document image degradation and in the context of lacking information on the document structure such as the document model and the typographical parameters. However, previous work in the use of texture analysis for segmentation of digitized historical document images has been limited to separately test one of the well-known texture-based approaches such as autocorrelation function, Grey Level Co-occurrence Matrix (GLCM), Gabor filters, gradient, wavelets, etc. In this paper we raise the question of which texture-based method could be better suited for discriminating on the one hand graphical regions from textual ones and on the other hand for separating textual regions with different sizes and fonts. The objective of this paper is to compare some of the well-known texture-based approaches: autocorrelation function, GLCM, and Gabor filters, used in a segmentation of digitized historical document images. Texture features are briefly described and quantitative results are obtained on simplified historical document images. The achieved results are very encouraging.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"10 1","pages":"102-109"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86967903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The HipPub Date : 2013-08-24DOI: 10.1145/2501115.2501117
Rafi Cohen, Abedelkadir Asi, K. Kedem, Jihad El-Sana, I. Dinstein
{"title":"Robust text and drawing segmentation algorithm for historical documents","authors":"Rafi Cohen, Abedelkadir Asi, K. Kedem, Jihad El-Sana, I. Dinstein","doi":"10.1145/2501115.2501117","DOIUrl":"https://doi.org/10.1145/2501115.2501117","url":null,"abstract":"We present a method to segment historical document images into regions of different content. First, we segment text elements from non-text elements using a binarized version of the document. Then, we refine the segmentation of the non-text regions into drawings, background and noise. At this stage, spatial and color features are exploited to guarantee coherent regions in the final segmentation. Experiments show that the suggested approach achieves better segmentation quality with respect to other methods. We examine the segmentation quality on 252 pages of a historical manuscript, for which the suggested method achieves about 92% and 90% segmentation accuracy of drawings and text elements, respectively.","PeriodicalId":77938,"journal":{"name":"The Hip","volume":"19 1","pages":"110-117"},"PeriodicalIF":0.0,"publicationDate":"2013-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76731288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}