{"title":"Detection of comic books twin pages with a non-overlapping stitching method","authors":"C. Guérin, J. Burie, J. Ogier","doi":"10.1145/3011549.3011550","DOIUrl":"https://doi.org/10.1145/3011549.3011550","url":null,"abstract":"We address in this paper the issue of stitching non overlapping pairs of real digitized comic books' pages. The main objective is to be able to decide if two pages are meant to be displayed together for their content to be understood. First, the relevant content is separated from the background so the pairing is done over relevant pieces of visual information. We define the different kinds of noise that one can found on the edges of such documents and how to get rid of it. Then we propose a method to decide whether a couple of pages should be paired or not, based on the analysis of the relevant content. Our compatibility methods are evaluated against the methods from the literature, adapted from jigsaw puzzles solvers. Results are discussed over an actual commercial dataset of digitized comic books.","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121221580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Retrieval of comic book images using context relevance information","authors":"Thanh Nam Le, M. Luqman, J. Burie, J. Ogier","doi":"10.1145/3011549.3011561","DOIUrl":"https://doi.org/10.1145/3011549.3011561","url":null,"abstract":"Despite the widespread research interest given in the recent years in analyzing the structure and content of comic books, the question of how to effectively query and retrieve comic images stays a challenge, due to the substantial differences between them and naturalistic images. In this paper, we present a scheme to represent the content in comic-page images using attributed region adjacency graphs. The frequent subgraphs are then mined, and we propose a similarity score for the graphs based on the overlap between them in terms of common component frequent subgraphs. We show that the relationship between the computed similarity score versus panel order can help locating and grouping panels with similar content, or to detect the changing between \"scenes\", which eventually help to retrieve more relevant results.","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133653069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yukihiro Moriyama, Byeongseon Park, Shinnosuke Iwaoki, Mitsunori Matsushita
{"title":"Designing a question-answering system for comic contents","authors":"Yukihiro Moriyama, Byeongseon Park, Shinnosuke Iwaoki, Mitsunori Matsushita","doi":"10.1145/3011549.3011554","DOIUrl":"https://doi.org/10.1145/3011549.3011554","url":null,"abstract":"The objective of our research is to create a question answering system for comics. Because a comic has multimodal contents, we have to answer questions about text as well as illustrations. This is different from the conventional question answering system. To solve this problem, in this study, we organized the information to be obtained from comic illustrations and examined the framework of question answering for this content. Then, we built a prototype system and examined the question answering system for comic contents.","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116483525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Manga generator, a future of interactive manga media: invited talk paper","authors":"Akihiko Shirai","doi":"10.1145/3011549.3015156","DOIUrl":"https://doi.org/10.1145/3011549.3015156","url":null,"abstract":"This article contributes to the realization of an immersive role playing manga generating advertising system named \"Manga Generator\". The project began as a student virtual reality project and had underwent various collaborations with professional manga artists and governmental agencies. It was exhibited at many international exhibitions and is currently a permanent exhibition in a museum. The latest version \"MGV\" is a new advertising entertainment system that enables up to two players to engage in role playing interactive manga while allowing other audiences to see advertisement video simultaneously using multiplex hidden image technology \"ExPixel\". In this research, we combined interactive comic generating system with full body motion detection and real time image processing technologies. It can attract non-audiences who are not interested in the video announcements thanks to players playing Manga Generator.","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130442431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Toward speech text recognition for comic books","authors":"Christophe Rigaud, S. Pal, J. Burie, J. Ogier","doi":"10.1145/3011549.3011557","DOIUrl":"https://doi.org/10.1145/3011549.3011557","url":null,"abstract":"Speech text in comic books is placed and written in a particular manner by the letterers which raises unusual challenges for text recognition. We first detail these challenges and present different approaches to solve them. We compare the performances of generic versus specifically trained OCR systems for typewritten and handwritten text lines from French comic books. This work is evaluated over a subset of public (eBDtheque) and private (Sequencity) datasets. We demonstrate that generic OCR systems perform best on typewritten-like and lowercase fonts while specifically trained OCR can be very powerful on skewed, uppercase and even cursive fonts.","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114339215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Azuma Fujimoto, Toru Ogawa, Kazuyoshi Yamamoto, Yusuke Matsui, T. Yamasaki, K. Aizawa
{"title":"Manga109 dataset and creation of metadata","authors":"Azuma Fujimoto, Toru Ogawa, Kazuyoshi Yamamoto, Yusuke Matsui, T. Yamasaki, K. Aizawa","doi":"10.1145/3011549.3011551","DOIUrl":"https://doi.org/10.1145/3011549.3011551","url":null,"abstract":"We have created Manga109, a dataset of a variety of 109 Japanese comic books publicly available for use for academic purposes. This dataset provides numerous comic images but lacks the annotations of elements in the comics that are necessary for use by machine learning algorithms or evaluation of methods. In this paper, we present our ongoing project to build metadata for Manga109. We first define the metadata in terms of frames, texts and characters. We then present our web-based software for efficiently creating the ground truth for these images. In addition, we provide a guideline for the annotation with the intent of improving the quality of the metadata.","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121289732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Comic visualization on smartphones based on eye tracking","authors":"Olivier Augereau, Mizuki Matsubara, K. Kise","doi":"10.1145/3011549.3011553","DOIUrl":"https://doi.org/10.1145/3011549.3011553","url":null,"abstract":"The visualization of comic images on a small screen is a difficult problem as the image is too large to be displayed on the screen and we do not know which areas and in which order the users want to see the image. The basic solution for the user is to look at the image in full screen without being able to see the details, or to zoom and scroll through the image, which can be quite inconvenient if the interactions have to often be repeated. Our idea is to use an eye tracker to record where the users reading a comic on paper books or large screens are looking at, to reproduce their reading behaviors with a comic visualization system and guide the users using a smaller screen through the comic.","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116884735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mizuki Matsubara, Olivier Augereau, Charles Lima Sanches, K. Kise
{"title":"Emotional arousal estimation while reading comics based on physiological signal analysis","authors":"Mizuki Matsubara, Olivier Augereau, Charles Lima Sanches, K. Kise","doi":"10.1145/3011549.3011556","DOIUrl":"https://doi.org/10.1145/3011549.3011556","url":null,"abstract":"Estimating emotions while reading enables new services such as comic recommendation. Most of existing emotion estimation systems employ bulky devices. Furthermore, few applications have been developed for analyzing the emotions while reading. The purpose of our research is to develop a method for estimating emotions while reading. As the target of reading, we select comics, which stimulate emotions often more than other types of documents. As we want our system to be easily usable, we selected sensors embedded in a wristband and an eye tracker. Emotions can be described by two dimensions called emotional valence and arousal. As a first step, we propose in this paper to estimate the emotional arousal. We analyze the electrodermal activity, blood volume pulse, heart rate, skin temperature and pupil diameter of a subject to estimate if the reader feels a high or low arousal while reading. Our experiment shows that for some participants, the arousal can be estimated accurately.","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128744430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Comics image processing: learning to segment text","authors":"N. Hirata, Igor dos Santos Montagner, R. Hirata","doi":"10.1145/3011549.3011560","DOIUrl":"https://doi.org/10.1145/3011549.3011560","url":null,"abstract":"We employ an image operator learning method to segment text in comic images. Since the method is based on learning from pairs of input and corresponding expected output images, it is flexible with respect to alphabet sets and text orientation. The method is applied on both Japanese and European comics. Results indicate that most text regions can be straightforwardly identified from the output images.","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126325108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. Ogier, K. Aizawa, K. Osaka, J. Burie, T. Yamasaki, Motoi Iwata Osaka
{"title":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","authors":"J. Ogier, K. Aizawa, K. Osaka, J. Burie, T. Yamasaki, Motoi Iwata Osaka","doi":"10.1145/3011549","DOIUrl":"https://doi.org/10.1145/3011549","url":null,"abstract":"","PeriodicalId":319382,"journal":{"name":"Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding","volume":"661 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131987857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}