Sabrina Kletz, Klaus Schöffmann, Bernd Münzer, Manfred Jürgen Primus, H. Husslein
{"title":"Surgical Action Retrieval for Assisting Video Review of Laparoscopic Skills","authors":"Sabrina Kletz, Klaus Schöffmann, Bernd Münzer, Manfred Jürgen Primus, H. Husslein","doi":"10.1145/3132390.3132395","DOIUrl":"https://doi.org/10.1145/3132390.3132395","url":null,"abstract":"An increasing number of surgeons promote video review of laparoscopic surgeries for detection of technical errors at an early stage as well as for training purposes. The reason behind is the fact that laparoscopic surgeries require specific psychomotor skills, which are difficult to learn and teach. The manual inspection of surgery video recordings is extremely cumbersome and time-consuming. Hence, there is a strong demand for automated video content analysis methods. In this work, we focus on retrieving surgical actions from video collections of gynecologic surgeries. We propose two novel dynamic content descriptors for similarity search and investigate a query-by-example approach to evaluate the descriptors on a manually annotated dataset consisting of 18 hours of video content. We compare several content descriptors including dynamic information of the segments as well as descriptors containing only spatial information of keyframes of the segments. The evaluation shows that our proposed dynamic content descriptors considering motion and spatial information from the segment achieve a better retrieval performance than static content descriptors ignoring temporal information of the segment at all. The proposed content descriptors in this work enable content-based video search for similar laparoscopic actions, which can be used to assist surgeons in evaluating laparoscopic surgical skills.","PeriodicalId":123540,"journal":{"name":"Proceedings of the 2017 ACM Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128565868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Houssem Chatbri, Kevin McGuinness, S. Little, Jiang Zhou, K. Kameyama, P. Kwan, N. O’Connor
{"title":"Automatic MOOC Video Classification using Transcript Features and Convolutional Neural Networks","authors":"Houssem Chatbri, Kevin McGuinness, S. Little, Jiang Zhou, K. Kameyama, P. Kwan, N. O’Connor","doi":"10.1145/3132390.3132393","DOIUrl":"https://doi.org/10.1145/3132390.3132393","url":null,"abstract":"The amount of MOOC video materials has grown exponentially in recent years. Therefore, their storage and analysis need to be made as fully automated as possible in order to maintain their management quality. In this work, we present a method for automatic topic classification of MOOC videos using speech transcripts and convolutional neural networks (CNN). Our method works as follows: First, speech recognition is used to generate video transcripts. Then, the transcripts are converted into images using a statistical co-occurrence transformation that we designed. Finally, a CNN is used to produce video category labels for a transcript image input. For our data, we use the Khan Academy on a Stick dataset that contains 2,545 videos, where each video is labeled with one or two of 13 categories. Experiments show that our method is strongly competitive against other methods that are also based on transcript features and supervised learning.","PeriodicalId":123540,"journal":{"name":"Proceedings of the 2017 ACM Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training","volume":"166 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131682595","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Wan-Lun Tsai, Ming-Fen Chung, Tse-Yu Pan, Min-Chun Hu
{"title":"Train in Virtual Court: Basketball Tactic Training via Virtual Reality","authors":"Wan-Lun Tsai, Ming-Fen Chung, Tse-Yu Pan, Min-Chun Hu","doi":"10.1145/3132390.3132394","DOIUrl":"https://doi.org/10.1145/3132390.3132394","url":null,"abstract":"In this paper, we present a basketball tactic training system based on multimedia and virtual reality (VR) technologies to improve the effectiveness and experience of tactic learning. A tablet-based digital tactic board (2D BTB) is developed to draw, select, or search for a target offensive tactic. The 2D player trajectories of the target tactic are then converted into 3D player animation immediately to provide virtual reality content for experiencing and practicing the target basketball tactic. Through the VR environment, the learner can vividly experience how a tactic is executed in a global view or from a specific player's viewing direction. The basketball tactic movement guidance and virtual defenders are rendered in our VR system according to the target offensive tactic and the learner's head pose. We also design an experimental process to validate that the proposed VR training system not only makes the learner feel more like playing in a real basketball game but also improves the efficiency and effectiveness of learning new basketball tactics.","PeriodicalId":123540,"journal":{"name":"Proceedings of the 2017 ACM Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134390790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Scherp, V. Mezaris, T. Köhler, Alexander Hauptmann
{"title":"Proceedings of the 2017 ACM Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training","authors":"A. Scherp, V. Mezaris, T. Köhler, Alexander Hauptmann","doi":"10.1145/3132390","DOIUrl":"https://doi.org/10.1145/3132390","url":null,"abstract":"It is our great pleasure to welcome you to the 1st International Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training -- MultiEdTech 2017. Educational and Knowledge Technologies (EdTech), especially in connection to multimedia content and the vision of mobile and personalized learning, is a hot topic in both academia and the business start-ups ecosystem. The driver and enabler of this is on the one side the development and widespread availability of multimedia materials and MOOCs, which represent multimedia content produced specifically for supporting e-learning; and, on the other side, the ever increasing availability of all sorts on information on the Internet and in social media channels (e. g., lectures, research papers, user-generated videos, news items), which, despite not directly targeting e-learning, can prove to be valuable complements to the more targeted learning materials. Although the availability of such content is not a problem these days, finding the right content and associating different relevant pieces of multimedia so as to enable a comprehensive learning experience on a chosen subject is by no means a trivial task. \u0000 \u0000The MultiEdTech 2017 workshop provides a forum for presenting research in areas related to multimedia-based educational and knowledge technologies and particularly on the use of multimedia search and retrieval, analysis and understanding, browsing, summarization, recommendation, and visualization technologies on multimedia content available in specialized learning platforms, the Web, mobile devices and/or social networks for supporting personalized and adaptive e-learning and training. \u0000 \u0000The workshop will be kicked off with an exciting keynote talk: \u0000Dr. Pablo Cesar from CWI, Amsterdam on \"Sensing Engagement: Helping Performers to Evaluate their Impact\". Dr. Cesar leads the Interactive and Distributed Systems group at CWI, which focuses on facilitating and improving the way people access media and communicate with others and with the environment. The keynote will overview on gathering data and understanding the experience of people attending cultural events, public lectures, and courses by using wearable sensor technology. Through practical case studies in different areas of the creative industries and education, it will showcase results and discuss about failures. \u0000 \u0000 \u0000 \u0000As paper presentations, we cover: \u0000\"Train in Virtual Court: Basketball Tactic Training via Virtual Reality\" by Wan-Lun Tsai, Ming-Fen Chung, Tse-Yu Pan, and Min-Chun Hu, presents a basketball tactic training system based on multimedia and virtual reality (VR) technologies. \u0000Sabrina Kletz, Klaus Schoeffmann, Bernd Munzer, and Manfred J. Primus present with \"Surgical Action Retrieval for Assisting Video Review of Laparoscopic Skills\" an information retrieval system to find surgical actions from video collections of gynecologic surgeries based on two novel content descriptors. \u0000Houssem C","PeriodicalId":123540,"journal":{"name":"Proceedings of the 2017 ACM Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124201755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Sensing Engagement: Helping Performers to Evaluate their Impact","authors":"Pablo César","doi":"10.1145/3132390.3132391","DOIUrl":"https://doi.org/10.1145/3132390.3132391","url":null,"abstract":"The keynote will provide an overview about different mechanisms to gather data by using wearable sensor technology for understanding the experience of people attending cultural events, public lectures, and courses. Through practical case studies in different areas of the creative industries and education, we will showcase our results and discuss about our failures. Based on realistic testing grounds, collaborating with several commercial and academic partners, we have deployed our technology and infrastructure in places such as the National Theatre of China in Shanghai. Our approach is to seamless connecting fashion and textiles with sensing technology, and with the environment. The final objective is to create intelligent and empathic systems that can react to the audience and their experience.","PeriodicalId":123540,"journal":{"name":"Proceedings of the 2017 ACM Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116444242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Chat2Doc: From Chats to How-to Instructions, FAQ, and Reports","authors":"Britta Meixner, Matthew L. Lee, S. Carter","doi":"10.1145/3132390.3132392","DOIUrl":"https://doi.org/10.1145/3132390.3132392","url":null,"abstract":"Sharing multimedia via messaging apps is widely used. However, the timeline structure makes it difficult to retrieve content shared over time. It is not possible to organize accumulated knowledge so that it is concise for future use and easy access. So far, no system exists that combines the easy-to-use interface of a messaging app with a knowledge extraction system that can create multimedia documents and allows users to store and edit content for future use. In this paper, we propose a system that will enable individuals to collect, store, and automatically extract procedural knowledge from their messaging interactions. The system uses the well-known chat interface to communicate and adds the capability for users to tag text and media to organize content. It also adds a new thread-like structure to the previously only linear timeline of a chat. Knowledge from the chat can then be extracted into a high-quality multimedia document.","PeriodicalId":123540,"journal":{"name":"Proceedings of the 2017 ACM Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training","volume":"154 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129469788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Session details: Oral Session","authors":"V. Mezaris","doi":"10.1145/3258508","DOIUrl":"https://doi.org/10.1145/3258508","url":null,"abstract":"","PeriodicalId":123540,"journal":{"name":"Proceedings of the 2017 ACM Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129975864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Session details: Keynote Address","authors":"V. Mezaris","doi":"10.1145/3258507","DOIUrl":"https://doi.org/10.1145/3258507","url":null,"abstract":"","PeriodicalId":123540,"journal":{"name":"Proceedings of the 2017 ACM Workshop on Multimedia-based Educational and Knowledge Technologies for Personalized and Social Online Training","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126813597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}