Manh-Khanh Ngo Huu, Sy-Tuyen Ho, Vinh-Tiep Nguyen, T. Ngo
{"title":"多语言gan:一种基于多语言gan的手写生成方法","authors":"Manh-Khanh Ngo Huu, Sy-Tuyen Ho, Vinh-Tiep Nguyen, T. Ngo","doi":"10.1109/MAPR53640.2021.9585285","DOIUrl":null,"url":null,"abstract":"Handwritten Text Recognition (HTR) is a difficult problem because of the diversity of calligraphic styles. To enhance the accuracy of HTR systems, a large amount of training data is required. The previous methods aim at generating handwritten images from input strings via RNN models such as LSTM or GRU. However, these methods require a predefined alphabet corresponding to a given language. Thus, they can not well adapt to a new languages. To address this problem, we propose an Image2Image-based method named Multilingual-GAN, which translates a printed text image into a handwritten style one. The main advantage of this approach is that the model does not depend on any language alphabets. Therefore, our model can be used on a new language without re-training on a new dataset. The quantitative results demonstrate that our proposed method outperforms other state-of-the-art models. Code is available at https://github.com/HoSyTuyen/MultilingualGAN","PeriodicalId":233540,"journal":{"name":"2021 International Conference on Multimedia Analysis and Pattern Recognition (MAPR)","volume":"79 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Multilingual-GAN: A Multilingual GAN-based Approach for Handwritten Generation\",\"authors\":\"Manh-Khanh Ngo Huu, Sy-Tuyen Ho, Vinh-Tiep Nguyen, T. Ngo\",\"doi\":\"10.1109/MAPR53640.2021.9585285\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Handwritten Text Recognition (HTR) is a difficult problem because of the diversity of calligraphic styles. To enhance the accuracy of HTR systems, a large amount of training data is required. The previous methods aim at generating handwritten images from input strings via RNN models such as LSTM or GRU. However, these methods require a predefined alphabet corresponding to a given language. Thus, they can not well adapt to a new languages. To address this problem, we propose an Image2Image-based method named Multilingual-GAN, which translates a printed text image into a handwritten style one. The main advantage of this approach is that the model does not depend on any language alphabets. Therefore, our model can be used on a new language without re-training on a new dataset. The quantitative results demonstrate that our proposed method outperforms other state-of-the-art models. Code is available at https://github.com/HoSyTuyen/MultilingualGAN\",\"PeriodicalId\":233540,\"journal\":{\"name\":\"2021 International Conference on Multimedia Analysis and Pattern Recognition (MAPR)\",\"volume\":\"79 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Conference on Multimedia Analysis and Pattern Recognition (MAPR)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/MAPR53640.2021.9585285\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Multimedia Analysis and Pattern Recognition (MAPR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MAPR53640.2021.9585285","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multilingual-GAN: A Multilingual GAN-based Approach for Handwritten Generation
Handwritten Text Recognition (HTR) is a difficult problem because of the diversity of calligraphic styles. To enhance the accuracy of HTR systems, a large amount of training data is required. The previous methods aim at generating handwritten images from input strings via RNN models such as LSTM or GRU. However, these methods require a predefined alphabet corresponding to a given language. Thus, they can not well adapt to a new languages. To address this problem, we propose an Image2Image-based method named Multilingual-GAN, which translates a printed text image into a handwritten style one. The main advantage of this approach is that the model does not depend on any language alphabets. Therefore, our model can be used on a new language without re-training on a new dataset. The quantitative results demonstrate that our proposed method outperforms other state-of-the-art models. Code is available at https://github.com/HoSyTuyen/MultilingualGAN