{"title":"The best angle correction of basketball shooting based on the fusion of time series features and dual CNN","authors":"Meicai Xiao","doi":"10.1016/j.eij.2024.100579","DOIUrl":null,"url":null,"abstract":"<div><div>The best shooting angle correction of basketball based on intelligent image analysis is an important branch of the development of intelligent sports. However, the current method is limited by the variability of the shape base, ignoring dynamic features and visual information, and there are some problems in the process of feature extraction and correction of related actions. This paper proposes a method to correct the best shooting angle of basketball based on the fusion of time series characteristics and dual CNN. Segmenting the shooting video, taking the video frame as the input of the key node extraction network of the shooting action, obtaining the video frame with the sequence information of the bone points, extracting the continuous T-frame video stack from it, and inputting it into the spatial context feature extraction network in the shooting posture prediction model based on dual stream CNN (MobileNet V3 network with multi-channel attention mechanism fusion module), extract the space context features of shooting posture; The superimposed optical flow graph of continuous video frames containing sequence information of bone points is input into the time convolution network (combined with Bi-LSTM network of multi-channel attention mechanism fusion module), extract the skeleton temporal sequence features during the shooting movement, using the spatial context features and skeleton temporal sequence features extracted from the feature fusion module, and realizing the prediction of shooting posture through Softmax according to the fusion results, calculate the shooting release speed under this attitude, solve the shooting release angle, and complete the correction of the best shooting release angle by comparing with the set conditions. The experimental results show that this method can achieve the best shooting angle correction, and the training learning rate is 0.2 × 10–3, training loss is about 0.05; MPJPE and MPJVE indicators are the lowest, and Top-1 indicators are the highest; The shooting percentage is about 95 %.</div></div>","PeriodicalId":56010,"journal":{"name":"Egyptian Informatics Journal","volume":"28 ","pages":"Article 100579"},"PeriodicalIF":5.0000,"publicationDate":"2024-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Egyptian Informatics Journal","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1110866524001427","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The best shooting angle correction of basketball based on intelligent image analysis is an important branch of the development of intelligent sports. However, the current method is limited by the variability of the shape base, ignoring dynamic features and visual information, and there are some problems in the process of feature extraction and correction of related actions. This paper proposes a method to correct the best shooting angle of basketball based on the fusion of time series characteristics and dual CNN. Segmenting the shooting video, taking the video frame as the input of the key node extraction network of the shooting action, obtaining the video frame with the sequence information of the bone points, extracting the continuous T-frame video stack from it, and inputting it into the spatial context feature extraction network in the shooting posture prediction model based on dual stream CNN (MobileNet V3 network with multi-channel attention mechanism fusion module), extract the space context features of shooting posture; The superimposed optical flow graph of continuous video frames containing sequence information of bone points is input into the time convolution network (combined with Bi-LSTM network of multi-channel attention mechanism fusion module), extract the skeleton temporal sequence features during the shooting movement, using the spatial context features and skeleton temporal sequence features extracted from the feature fusion module, and realizing the prediction of shooting posture through Softmax according to the fusion results, calculate the shooting release speed under this attitude, solve the shooting release angle, and complete the correction of the best shooting release angle by comparing with the set conditions. The experimental results show that this method can achieve the best shooting angle correction, and the training learning rate is 0.2 × 10–3, training loss is about 0.05; MPJPE and MPJVE indicators are the lowest, and Top-1 indicators are the highest; The shooting percentage is about 95 %.
期刊介绍:
The Egyptian Informatics Journal is published by the Faculty of Computers and Artificial Intelligence, Cairo University. This Journal provides a forum for the state-of-the-art research and development in the fields of computing, including computer sciences, information technologies, information systems, operations research and decision support. Innovative and not-previously-published work in subjects covered by the Journal is encouraged to be submitted, whether from academic, research or commercial sources.