Manh-Khanh Ngo Huu, Sy-Tuyen Ho, Vinh-Tiep Nguyen, T. Ngo
{"title":"Multilingual-GAN: A Multilingual GAN-based Approach for Handwritten Generation","authors":"Manh-Khanh Ngo Huu, Sy-Tuyen Ho, Vinh-Tiep Nguyen, T. Ngo","doi":"10.1109/MAPR53640.2021.9585285","DOIUrl":null,"url":null,"abstract":"Handwritten Text Recognition (HTR) is a difficult problem because of the diversity of calligraphic styles. To enhance the accuracy of HTR systems, a large amount of training data is required. The previous methods aim at generating handwritten images from input strings via RNN models such as LSTM or GRU. However, these methods require a predefined alphabet corresponding to a given language. Thus, they can not well adapt to a new languages. To address this problem, we propose an Image2Image-based method named Multilingual-GAN, which translates a printed text image into a handwritten style one. The main advantage of this approach is that the model does not depend on any language alphabets. Therefore, our model can be used on a new language without re-training on a new dataset. The quantitative results demonstrate that our proposed method outperforms other state-of-the-art models. Code is available at https://github.com/HoSyTuyen/MultilingualGAN","PeriodicalId":233540,"journal":{"name":"2021 International Conference on Multimedia Analysis and Pattern Recognition (MAPR)","volume":"79 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Multimedia Analysis and Pattern Recognition (MAPR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MAPR53640.2021.9585285","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Handwritten Text Recognition (HTR) is a difficult problem because of the diversity of calligraphic styles. To enhance the accuracy of HTR systems, a large amount of training data is required. The previous methods aim at generating handwritten images from input strings via RNN models such as LSTM or GRU. However, these methods require a predefined alphabet corresponding to a given language. Thus, they can not well adapt to a new languages. To address this problem, we propose an Image2Image-based method named Multilingual-GAN, which translates a printed text image into a handwritten style one. The main advantage of this approach is that the model does not depend on any language alphabets. Therefore, our model can be used on a new language without re-training on a new dataset. The quantitative results demonstrate that our proposed method outperforms other state-of-the-art models. Code is available at https://github.com/HoSyTuyen/MultilingualGAN