Cristiane Ferreira, C. Salles, Luís Santos, Fernando A. M. Trinta, Windson Viana
{"title":"Towards a Model and a Textual Representation for Location-based Games","authors":"Cristiane Ferreira, C. Salles, Luís Santos, Fernando A. M. Trinta, Windson Viana","doi":"10.1145/3103010.3121035","DOIUrl":"https://doi.org/10.1145/3103010.3121035","url":null,"abstract":"Location-Based Mobile Games (LBMGs) are a subclass of pervasive games that make use of location technologies to consider the players' geographic position in the game rules and mechanics. This research presents LEGaL, a language to model and represent the structure and multimedia contents (e.g., video, audio, 3D objects, etc.) of LBMGs. LEGaL is an extension of NCL (Nested Context Language) that allows the modelling and representation of mission-based games by supporting spatial and temporal relationships between game elements.","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121069047","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
B. Gatos, G. Louloudis, N. Stamatopoulos, Giorgos Sfikas
{"title":"Historical Document Processing","authors":"B. Gatos, G. Louloudis, N. Stamatopoulos, Giorgos Sfikas","doi":"10.1145/3103010.3103026","DOIUrl":"https://doi.org/10.1145/3103010.3103026","url":null,"abstract":"This tutorial focuses on recent advances and ongoing developments for historical document processing. It includes the main challenges involved, the different tasks that have to be implemented as well as practices and technologies that currently exist in the literature. The focus is given on the most promising techniques, related projects as well as on existing datasets and competitions that can be proved useful to historical document processing research.","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122871129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatic Knowledge Base Construction from Scholarly Documents","authors":"Rabah A. Al-Zaidy, C. Lee Giles","doi":"10.1145/3103010.3121043","DOIUrl":"https://doi.org/10.1145/3103010.3121043","url":null,"abstract":"The continuing growth of published scholarly content on the web ensures the availability of the most recent scientific findings to researchers. Scholarly documents, such as research articles, are easily accessed by using academic search engines that are built on large repositories of scholarly documents. Scientific information extraction from documents into a structured knowledge graph representation facilitates automated machine understanding of a document's content. Traditional information extraction approaches, that either require training samples or a preexisting knowledge base to assist in the extraction, can be challenging when applied to large repositories of digital documents. Labeled training examples for such large scale are difficult to obtain for such datasets. Also, most available knowledge bases are built from web data and do not have sufficient coverage to include concepts found in scientific articles. In this paper we aim to construct a knowledge graph from scholarly documents while addressing both these issues. We propose a fully automatic, unsupervised system for scientific information extraction that does not build on an existing knowledge base and avoids manually-tagged training data. We describe and evaluate a constructed taxonomy that contains over 15k entities resulting from applying our approach to 10k documents.","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114468092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
V. Lombardo, R. Damiano, Antonio Pizzo, Carmi Terzulli
{"title":"The Intangible Nature of Drama Documents: an FRBR View","authors":"V. Lombardo, R. Damiano, Antonio Pizzo, Carmi Terzulli","doi":"10.1145/3103010.3103019","DOIUrl":"https://doi.org/10.1145/3103010.3103019","url":null,"abstract":"As a pervasive form of artistic expression through ages and me- dia, drama features a twofold nature of its tangible manifestations (theatrical performances, movies, books, etc.) and its intangible abstraction (the story of Cinderella underlying Disney movie and Perrault's fable). The encoding of the intangible drama abstraction of drama documents is relevant for the preservation of cultural heritage and the didactics and research on drama documents. This paper addresses the task of encoding the notion of intangible story abstraction from the drama documents. The reference model is provided by a computational ontology that formally encodes the elements that characterize a drama, for purposes of semantic link- ing and inclusion in annotation schemata. By providing a formal expression posited between drama as work and its manifestations, the ontology-based representation is compliant with the model of Functional Requirements for Bibliographic Records (FRBR).","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114650671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sagnik Das, Gaurav Mishra, A. Sudharshana, Roy Shilkrot
{"title":"The Common Fold: Utilizing the Four-Fold to Dewarp Printed Documents from a Single Image","authors":"Sagnik Das, Gaurav Mishra, A. Sudharshana, Roy Shilkrot","doi":"10.1145/3103010.3121030","DOIUrl":"https://doi.org/10.1145/3103010.3121030","url":null,"abstract":"Handheld cameras are currently the device of choice for performing document digitization, due to their convenience, ubiquity and high performance at low cost. Software methods process a captured image, to rectify distortions and reconstruct the original document. Existing methods struggle to reconstruct a flattened version given a single image of a document distorted by folding. We propose a novel non-parametric page dewarping approach from a single image based on deep learning to identify creases due to folds on the paper. Our method then performs a 2D boundary method based on polynomial regression, and a Coons patch, to get a flattened reconstruction. We found our method improves OCR word accuracy by more than 2.5 times when compared to the original distorted image.","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122686019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Carlos Badenes-Olmedo, José Luis Redondo García, Óscar Corcho
{"title":"Distributing Text Mining tasks with librAIry","authors":"Carlos Badenes-Olmedo, José Luis Redondo García, Óscar Corcho","doi":"10.1145/3103010.3121040","DOIUrl":"https://doi.org/10.1145/3103010.3121040","url":null,"abstract":"We present librAIry, a novel architecture to store, process and analyze large collections of textual resources, integrating existing algorithms and tools into a common, distributed, high-performance workflow. Available text mining techniques can be incorporated as independent plug&play modules working in a collaborative manner into the framework. In the absence of a pre-defined flow, librAIry leverages on the aggregation of operations executed by different components in response to an emergent chain of events. Extensive use of Linked Data (LD) and Representational State Transfer (REST) principles are made to provide individually addressable resources from textual documents. We have described the architecture design and its implementation and tested its effectiveness in real-world scenarios such as collections of research papers, patents or ICT aids, with the objective of providing solutions for decision makers and experts in those domains. Major advantages of the framework and lessons-learned from these experiments are reported.","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"1999 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131263435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Using Abstract Anchors to Aid The Development of Multimedia Applications With Sensory Effects","authors":"R. Abreu, J. Santos","doi":"10.1145/3103010.3103014","DOIUrl":"https://doi.org/10.1145/3103010.3103014","url":null,"abstract":"Declarative multimedia authoring languages allows authors to combine multiple media objects, generating a range of multimedia presentations. Novel multimedia applications, focusing at improving user experience, extend multimedia applications with multisensory content. The idea is to synchronize sensory effects with the audiovisual content being presented. The usual approach for specifying such synchronization is to mark the content of a main media object (e.g. a main video) indicating the moments when a given effect has to be executed. For example, a mark may represent when snow appears in the main video so that a cold wind may be synchronized with it. Declarative multimedia authoring languages provide a way to mark subparts of a media object through anchors. An anchor indicates its begin and end times (video frames or audio samples) in relation to its parent media object. The manual definition of anchors in the above scenario is both not efficient and error prone (i) when the main media object size increases, (ii) when a given scene component appears several times and (iii) when the application requires marking scene components. This paper tackles this problem by providing an approach for creating abstract anchors in declarative multimedia documents. An abstract anchor represents (possibly) several media anchors, indicating the moments when a given scene component appears in a media object content. The author, therefore is able to define the application behavior through relationships among, for example, sensory effects and abstract anchors. Prior to executing, abstract anchors are automatically instantiated for each moment a given element appears and relationships are cloned so the application behavior is maintained. This paper presents an implementation of the proposed approach using NCL (Nested Context Language) as the target language. The abstract anchor processor is implemented in Lua and uses available APIs for video recognition in order to identify the begin and end times for abstract anchor instances. We also present an evaluation of our approach using a real world use cases.","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115562002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
C. C. Viel, Kamila R. H. Rodrigues, Isabela Zaine, Bruna C. R. Cunha, L. Scalco, M. G. Pimentel
{"title":"Personalized Ubiquitous Data Collection and Intervention as Interactive Multimedia Documents","authors":"C. C. Viel, Kamila R. H. Rodrigues, Isabela Zaine, Bruna C. R. Cunha, L. Scalco, M. G. Pimentel","doi":"10.1145/3103010.3121046","DOIUrl":"https://doi.org/10.1145/3103010.3121046","url":null,"abstract":"The Experience Sampling Method (ESM) has been proposed as a method for collecting data about people's experiences in their everyday and natural environments. ESM-based systems offer limited authoring for interactive documents designed to collect text-based responses offered as answers to text-based questions, and integrated with the non-intrusive data collection from sensors. From a document engineering perspective, ESM brings new requirements with respect to the authoring of non-trivial interaction and navigation workflow, in particular when multiple media and collaborative tasks are concerned. Tackling existing challenges, we modeled the Experience Sampling and Programmed Intervention Method (ESPIM) by combining ESM, individualized teaching procedures and ubiquitous computing toward producing interactive personalized multimedia documents applied in data collection.","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114724602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jie Mei, Aminul Islam, A. Mohammad, Yajing Wu, E. Milios
{"title":"Post-Processing OCR Text using Web-Scale Corpora","authors":"Jie Mei, Aminul Islam, A. Mohammad, Yajing Wu, E. Milios","doi":"10.1145/3103010.3121032","DOIUrl":"https://doi.org/10.1145/3103010.3121032","url":null,"abstract":"We introduce a (semi-)automatic OCR post-processing system that utilizes web-scale linguistic corpora in providing high-quality correction. This paper is a comprehensive system overview with the focus on the computational procedures, applied linguistic analysis, and processing optimization.","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"23 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123438406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. F. Orlando, Isabela Zaine, M. G. Pimentel, D. G. Souza, C. Teixeira
{"title":"Interactive Documents based on Discrete Trials","authors":"A. F. Orlando, Isabela Zaine, M. G. Pimentel, D. G. Souza, C. Teixeira","doi":"10.1145/3103010.3121048","DOIUrl":"https://doi.org/10.1145/3103010.3121048","url":null,"abstract":"Interactive documents offer users alternatives for accessing the content available. In Education, researchers employ Individualized Learning Programs as interactive multimedia documents in order to teach students a variety of subjects. In order to author such interactive document, domain experts have to control features such as pace, duration, response-based criteria, and the hierarchy of learning units. We present how individualized learning programs can be modeled as interactive documents based on discrete trials and deterministic finite automata. We also report the main numbers associated with the use of a companion system deployed in real environments by domain specialists and learners.","PeriodicalId":200469,"journal":{"name":"Proceedings of the 2017 ACM Symposium on Document Engineering","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126272530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}