MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355520
Aaron K. Baughman, Eduardo Morales, G. Reiss, Nancy Greco, Stephen Hammer, Shiqiang Wang
{"title":"Detection of Tennis Events from Acoustic Data","authors":"Aaron K. Baughman, Eduardo Morales, G. Reiss, Nancy Greco, Stephen Hammer, Shiqiang Wang","doi":"10.1145/3347318.3355520","DOIUrl":"https://doi.org/10.1145/3347318.3355520","url":null,"abstract":"Professional tennis is a fast-paced sport with serves and hits that can reach speeds of over 100 mph and matches lasting long in duration. For example, in 13 years of Grand Slam data, there were 454 matches with an average of 3 sets that lasted 40 minutes. The fast pace and long duration of tennis matches make tracking the time boundaries of each tennis point in a match challenging. The visual aspect of a tennis match is highly diverse because of its variety in angles, occlusions, resolutions, contrast and colors, but the sound component is relatively stable and consistent. In this paper, we present a system that detects events such as ball hits and point boundaries in a tennis match from sound data recorded in the match. We first describe the sound processing pipeline that includes preprocessing, feature extraction, basic (atomic) event detection, and point boundary detection. Then, we describe the overall cloud-based system architecture. Afterwards, we describe the user interface that includes a tool for data labeling to efficiently generate the training dataset, and a workbench for sound and model management. The performance of our system is evaluated in experiments with real-world tennis sound data. Our proposed pipeline can detect atomic tennis events with an F1-score of 92.39% and point boundaries with average precision and recall values of around 80%. This system can be very useful for tennis coaches and players to find and extract game highlights with specific characteristics, so that they can analyze these highlights and establish their play strategy.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127205146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355519
Philipp Seidenschwarz, Adalsteinn Jonsson, Fabian Rauschenbach, M. Rumo, L. Probst, H. Schuldt
{"title":"Combining Qualitative and Quantitative Analysis in Football with SportSense","authors":"Philipp Seidenschwarz, Adalsteinn Jonsson, Fabian Rauschenbach, M. Rumo, L. Probst, H. Schuldt","doi":"10.1145/3347318.3355519","DOIUrl":"https://doi.org/10.1145/3347318.3355519","url":null,"abstract":"The task of performance analysts and coaches in football (and other team sports) is manifold: they need to assess the performance of individual players of their team, they need to monitor the interaction between players of their team and their tactical compliance, and they need to analyze other teams. For this, they usually have to consider various sources of information: video footage, tracking data, event data, and aggregated statistics. On the basis of this information, analysts have to generate quantitative summaries of events including their spatial and temporal distribution, and the qualitative assessment of individual events by considering the associated video footage. In this paper, we present SportSense, a system for sports video retrieval, that seamlessly combines quantitative and qualitative analysis. For this, SportSense provides dedicated filters that help analysts in selecting the events they are interested in. Moreover, it supports the comparative analysis of stored queries with respect to specific parameters. Essentially, SportSense allows to easily switch between qualitative and quantitative analyses to support coaches and analysts in a best possible way in their task. Based on a user study, we show the effectiveness of the proposed approach.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133879407","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355522
D. Saupe, Alexander Artiga Gonzalez, Ramona Burger, C. Abbiss
{"title":"Empirical Analysis of Pacing in Road Cycling","authors":"D. Saupe, Alexander Artiga Gonzalez, Ramona Burger, C. Abbiss","doi":"10.1145/3347318.3355522","DOIUrl":"https://doi.org/10.1145/3347318.3355522","url":null,"abstract":"The pacing profile adopted throughout a competitive time trial may be decisive in the overall outcomes of the event. Riders distribute their energy resources based on a range of factors including prior experience, perception of effort, knowledge of distance to cover and potential motivation. Some athletes and professional cycling teams may also quantify individual pacing strategies derived from computational scientific methods. In this work we collect and analyze data of self-selected individual pacing profiles from approximately 12,000 competitive riders on a well-known hill climbing road segment in the Adelaide Hills, South Australia. We found that riders chose from a variety of very different pacing profiles, including some opposing profiles. For the classification of pacing this paper describes the pipeline of collection GPS-based and time stamped performance data, data filtering, augmentation of road gradient and power values, and the classification procedure.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121617422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355516
T. Haruyama, Sho Takahashi, Takahiro Ogawa, M. Haseyama
{"title":"Retrieval of Similar Scenes Based on Multimodal Distance Metric Learning in Soccer Videos","authors":"T. Haruyama, Sho Takahashi, Takahiro Ogawa, M. Haseyama","doi":"10.1145/3347318.3355516","DOIUrl":"https://doi.org/10.1145/3347318.3355516","url":null,"abstract":"This paper presents a new method for retrieval of similar scenes based on multimodal distance metric learning in far-view soccer videos that broadly capture soccer fields and are not edited. We extract visual features and audio features from soccer video clips, and we extract text features from text data corresponding to these soccer video clips. In addition, distance metric learning based on Laplacian Regularized Metric Learning is performed to calculate the distances for each kind of features. Finally, by determining the final rank by integrating these distances, we realize successful multimodal retrieval of similar scenes from query scenes of soccer video clips. Experimental results show the effectiveness of our retrieval method.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122858420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355521
K. Abe, Chikara Nakamura, Yosuke Otsubo, Tetsuya Koike, N. Yokoya
{"title":"Spectator Excitement Detection in Small-scale Sports Events","authors":"K. Abe, Chikara Nakamura, Yosuke Otsubo, Tetsuya Koike, N. Yokoya","doi":"10.1145/3347318.3355521","DOIUrl":"https://doi.org/10.1145/3347318.3355521","url":null,"abstract":"Detection of the excitement of spectators in sports is useful for various applications such as automatic highlight generation and automatic video editing. Therefore, spectator analysis has been widely studied. The two main approaches used for this include holistic and object-based approaches. Holistic approaches have been applied in most previous works, however, they do not work in small-scale games, where there are fewer spectators compared to those of large-scale games. In this work, we propose a method for detecting the state of excitement of spectators in small-scale games using an object-based approach. To evaluate our method, we build our own datasets consisting of both spectator and player videos. Experimental results show that our method outperforms a holistic baseline method and allows excitement detection of individual spectators.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132006803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355525
Moritz Einfalt, Charles Dampeyrou, D. Zecha, R. Lienhart
{"title":"Frame-Level Event Detection in Athletics Videos with Pose-Based Convolutional Sequence Networks","authors":"Moritz Einfalt, Charles Dampeyrou, D. Zecha, R. Lienhart","doi":"10.1145/3347318.3355525","DOIUrl":"https://doi.org/10.1145/3347318.3355525","url":null,"abstract":"In this paper we address the problem of automatic event detection in athlete motion for automated performance analysis in athletics. We specifically consider the detection of stride-, jump- and landing related events from monocular recordings in long and triple jump. Existing work on event detection in sports often uses manually designed features on body and pose configurations of the athlete to infer the occurrence of events. We present a two-step approach, where temporal 2D pose sequences extracted from the videos form the basis for learning an event detection model. We formulate the detection of discrete events as a sequence translation task and propose a convolutional sequence network that can accurately predict the timing of event occurrences. Our best performing architecture achieves a precision/recall of 92.3%/89.0% in detecting start and end of ground contact during the run-up and jump of an athlete at a temporal precision of +/- 1 frame at 200Hz. The results show that 2D pose sequences are a suitable motion representation for learning event detection in a sequence-to-sequence framework.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128423102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355523
Tomohiro Shimizu, Ryo Hachiuma, H. Saito, Takashi Yoshikawa, Chonho Lee
{"title":"Prediction of Future Shot Direction using Pose and Position of Tennis Player","authors":"Tomohiro Shimizu, Ryo Hachiuma, H. Saito, Takashi Yoshikawa, Chonho Lee","doi":"10.1145/3347318.3355523","DOIUrl":"https://doi.org/10.1145/3347318.3355523","url":null,"abstract":"In this paper, we propose a method to predict the future shot direction in a tennis match using pose information and player position. As far as we know, there is no work that deals with such a predictive task, so there is no shot direction dataset as yet. Therefore, using a YouTube tennis match video, we construct an time of impact and shot direction dataset. To reduce annotation costs, we propose a method to automatically label the shot direction. Moreover, we propose a method to predict the future shot direction using the constructed dataset. The shot direction is predicted using LSTM(long short-time memory), from sequential pose information up to the time of impact and the player position. We employ OpenPose to extract the position of skeleton joints. In the experiment, we evaluate the accuracy of shot direction prediction and verify the effectiveness of the proposed method. Since there are no studies that predict future shot direction, we set four baseline methods to evaluate the effectiveness of our proposed method.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121157574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355526
F. Turchini, Lorenzo Seidenari, L. Galteri, Andrea Ferracani, Giuseppe Becchi, A. Bimbo
{"title":"Flexible Automatic Football Filming and Summarization","authors":"F. Turchini, Lorenzo Seidenari, L. Galteri, Andrea Ferracani, Giuseppe Becchi, A. Bimbo","doi":"10.1145/3347318.3355526","DOIUrl":"https://doi.org/10.1145/3347318.3355526","url":null,"abstract":"We propose a method aimed at reducing human intervention in football video shooting and highlights editing, allowing automatic highlight detection together with panning and zooming on salient areas of the playing field. Our recognition subsystem exploits computer vision algorithms to perform automatic detection, pan and zoom and extraction of salient segments of a recorded match. Matches are elaborated offline, extracting and analyzing motion and visual features of the elements in salient zones of the scene, i.e. midfield circle and penalty areas. Automatic summarization is performed by classifying subsequences of a match with machine learning algorithms, which are pretrained on previously acquired and annotated videos of other matches. Among salient actions, special attention is given to goal events, but also other generic highlights are identified. The only assumption for our method to work is to employ a pair of cameras which should frame the football pitch splitting the field in two halves. We demonstrate the functioning of our approach using two ultra high definition cameras, building a system which is also able to collect various metadata of the matches to extrapolate other salient information.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125027282","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355524
Melissa Sanabria, Sherly, F. Precioso, Thomas Menguy
{"title":"A Deep Architecture for Multimodal Summarization of Soccer Games","authors":"Melissa Sanabria, Sherly, F. Precioso, Thomas Menguy","doi":"10.1145/3347318.3355524","DOIUrl":"https://doi.org/10.1145/3347318.3355524","url":null,"abstract":"The massive growth of sports videos, specially in soccer, has resulted in a need for the automatic generation of summaries, where the objective is not only to show the most important actions of the match but also to elicit as much emotion as the ones bring upon by human editors. State-of-the-art methods on video summarization mostly rely on video processing, however this is not an optimal approach for long videos such as soccer matches. In this paper we propose a multimodal approach to automatically generate summaries of soccer match videos that consider both event and audio features. The event features get a shorter and better representation of the match, and the audio helps detect the excitement generated by the game. Our method consists of three consecutive stages: Proposals, Summarization and Content Refinement. The first one generates summary proposals, using Multiple Instance Learning to deal with the similarity between the events inside the summary and the rest of the match. The Summarization stage uses event and audio features as input of a hierarchical Recurrent Neural Network to decide which proposals should indeed be in the summary. And the last stage, takes advantage of the visual content to create the final summary. The results show that our approach outperforms by a large margin not only the video processing methods but also methods that use event and audio features.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133402517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MMSports '19Pub Date : 2019-10-15DOI: 10.1145/3347318.3355518
Mohammad Hedayati, M. Cree, Jonathan B. Scott
{"title":"Tracking Jockeys in a Cluttered Environment with Group Dynamics","authors":"Mohammad Hedayati, M. Cree, Jonathan B. Scott","doi":"10.1145/3347318.3355518","DOIUrl":"https://doi.org/10.1145/3347318.3355518","url":null,"abstract":"This project aims to detect and track jockeys at the turning point of the horse races. The detection and tracking of the objects is a very challenging task in a crowded environment such as horse racing due to occlusion. However, in the horse race, the jockeys follow each other's paths and move as a slowly changing group. This group dynamic gives an important cue to approximate the location of obscured jockeys. This paper proposes a novel approach to handle occlusion by the integration of the group dynamic into jockeys tracking framework. The experimental result shows the effect of group dynamics on the tracking performance against partial and full occlusions.","PeriodicalId":322390,"journal":{"name":"MMSports '19","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128529087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}