{"title":"基于细粒度文本到图像合成的知识图属性丰富","authors":"Yonghua Zhu, Ning Ge, Jieyu Huang, Yunwen Zhu, Binghui Zheng, Wenjun Zhang","doi":"10.1145/3487075.3487155","DOIUrl":null,"url":null,"abstract":"In this paper, we propose an Attribute-Rich Generative Adversarial Network (AttRiGAN) for text-to-image synthesis, which enriches the simple text description by associating knowledge graph and embedding it in the synthesis task in the form of an attribute matrix. Higher fine-grained images can be synthesized with AttRiGAN, and the synthesized sample are more similar to the objects that exist in the real world, since they are driven by attributes which are enriched from the knowledge graph. The experiments conducted on two widely-used fine-grained image datasets show that our AttRiGAN allows a significant improvement in fine-grained text-to-image synthesis.","PeriodicalId":354966,"journal":{"name":"Proceedings of the 5th International Conference on Computer Science and Application Engineering","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enriching Attributes from Knowledge Graph for Fine-grained Text-to-Image Synthesis\",\"authors\":\"Yonghua Zhu, Ning Ge, Jieyu Huang, Yunwen Zhu, Binghui Zheng, Wenjun Zhang\",\"doi\":\"10.1145/3487075.3487155\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we propose an Attribute-Rich Generative Adversarial Network (AttRiGAN) for text-to-image synthesis, which enriches the simple text description by associating knowledge graph and embedding it in the synthesis task in the form of an attribute matrix. Higher fine-grained images can be synthesized with AttRiGAN, and the synthesized sample are more similar to the objects that exist in the real world, since they are driven by attributes which are enriched from the knowledge graph. The experiments conducted on two widely-used fine-grained image datasets show that our AttRiGAN allows a significant improvement in fine-grained text-to-image synthesis.\",\"PeriodicalId\":354966,\"journal\":{\"name\":\"Proceedings of the 5th International Conference on Computer Science and Application Engineering\",\"volume\":\"7 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-10-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 5th International Conference on Computer Science and Application Engineering\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3487075.3487155\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 5th International Conference on Computer Science and Application Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3487075.3487155","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Enriching Attributes from Knowledge Graph for Fine-grained Text-to-Image Synthesis
In this paper, we propose an Attribute-Rich Generative Adversarial Network (AttRiGAN) for text-to-image synthesis, which enriches the simple text description by associating knowledge graph and embedding it in the synthesis task in the form of an attribute matrix. Higher fine-grained images can be synthesized with AttRiGAN, and the synthesized sample are more similar to the objects that exist in the real world, since they are driven by attributes which are enriched from the knowledge graph. The experiments conducted on two widely-used fine-grained image datasets show that our AttRiGAN allows a significant improvement in fine-grained text-to-image synthesis.