{"title":"Design of Image Generation System for DCGAN-Based Kids' Book Text","authors":"Jaehyeong Cho, Nammee Moon","doi":"10.3745/JIPS.02.0149","DOIUrl":null,"url":null,"abstract":"When a picture book is photographed with a smart device, the text is analyzed for meaning and associated images are created. Image creation is the first step in learning DCGAN using class lists and images. In this study, DCGAN was trained with 11 classes and images of 1688 bears, which were collected by ImageNet for design. The second step is to shoot the image and text of the picture book on a smart device, and convert the text part of the shot image into a system readable character. We use the morpheme analyzer to classify nouns and verbs in text, and Discriminator learn to recognize the classified parts of speech as latent vectors of images. The third step is to create an associated image in the text. In the picture book, take the text of the part without the image and extract nouns and verbs. The extracted parts of speech and the learned latent vector are used as Generator parameters to generate images associated with the text.","PeriodicalId":415161,"journal":{"name":"J. Inf. Process. Syst.","volume":"38 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"J. Inf. Process. Syst.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3745/JIPS.02.0149","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
When a picture book is photographed with a smart device, the text is analyzed for meaning and associated images are created. Image creation is the first step in learning DCGAN using class lists and images. In this study, DCGAN was trained with 11 classes and images of 1688 bears, which were collected by ImageNet for design. The second step is to shoot the image and text of the picture book on a smart device, and convert the text part of the shot image into a system readable character. We use the morpheme analyzer to classify nouns and verbs in text, and Discriminator learn to recognize the classified parts of speech as latent vectors of images. The third step is to create an associated image in the text. In the picture book, take the text of the part without the image and extract nouns and verbs. The extracted parts of speech and the learned latent vector are used as Generator parameters to generate images associated with the text.