Prior Knowledge-guided Hierarchical Action Quality Assessment with 3D Convolution and Attention Mechanism

IF 4.6 Q1 OPTICS
Haoyang Zhou, Teng Hou, Jitao Li
{"title":"Prior Knowledge-guided Hierarchical Action Quality Assessment with 3D Convolution and Attention Mechanism","authors":"Haoyang Zhou, Teng Hou, Jitao Li","doi":"10.1088/1742-6596/2632/1/012027","DOIUrl":null,"url":null,"abstract":"Abstract Recently, there has been a growing interest in the field of computer vision and deep learning regarding a newly emerging problem known as action quality assessment (AQA). However, most researchers still rely on the traditional approach of using models from the video action recognition field. Unfortunately, this approach overlooks crucial features in AQA, such as movement fluency and degree of completion. Alternatively, some researchers have employed the transformer paradigm to capture action details and overall action integrity, but the high computational cost associated with transformers makes them impractical for real-time tasks. Due to the diversity of action types, it is challenging to rely solely on a shared model for quality assessment of various types of actions. To address these issues, we propose a novel network structure for AQA, which is the first to integrate multi-model capabilities through a classification model. Specifically, we utilize a pre-trained I3D model equipped with a self-attention block for classification. This allows us to evaluate various categories of actions using just one model. Furthermore, we introduce self-attention mechanisms and multi-head attention into the traditional convolutional neural network. By systematically replacing the last few layers of the conventional convolutional network, our model gains a greater ability to sense the global coordination of different actions. We have verified the effectiveness of our approach on the AQA-7 dataset. In comparison to other popular models, our model achieves satisfactory performance while maintaining a low computational cost.","PeriodicalId":44008,"journal":{"name":"Journal of Physics-Photonics","volume":"97 2","pages":"0"},"PeriodicalIF":4.6000,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Physics-Photonics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1088/1742-6596/2632/1/012027","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"OPTICS","Score":null,"Total":0}
引用次数: 0

Abstract

Abstract Recently, there has been a growing interest in the field of computer vision and deep learning regarding a newly emerging problem known as action quality assessment (AQA). However, most researchers still rely on the traditional approach of using models from the video action recognition field. Unfortunately, this approach overlooks crucial features in AQA, such as movement fluency and degree of completion. Alternatively, some researchers have employed the transformer paradigm to capture action details and overall action integrity, but the high computational cost associated with transformers makes them impractical for real-time tasks. Due to the diversity of action types, it is challenging to rely solely on a shared model for quality assessment of various types of actions. To address these issues, we propose a novel network structure for AQA, which is the first to integrate multi-model capabilities through a classification model. Specifically, we utilize a pre-trained I3D model equipped with a self-attention block for classification. This allows us to evaluate various categories of actions using just one model. Furthermore, we introduce self-attention mechanisms and multi-head attention into the traditional convolutional neural network. By systematically replacing the last few layers of the conventional convolutional network, our model gains a greater ability to sense the global coordination of different actions. We have verified the effectiveness of our approach on the AQA-7 dataset. In comparison to other popular models, our model achieves satisfactory performance while maintaining a low computational cost.
基于三维卷积和注意机制的先验知识引导分层动作质量评估
最近,在计算机视觉和深度学习领域中出现了一个新兴的问题——动作质量评估(AQA),引起了人们越来越多的兴趣。然而,大多数研究人员仍然依赖于传统的方法,使用视频动作识别领域的模型。不幸的是,这种方法忽略了AQA中的关键功能,如动作流畅性和完成度。另外,一些研究人员已经采用变压器范例来捕获动作细节和整体动作完整性,但是变压器相关的高计算成本使得它们不适合实时任务。由于行动类型的多样性,仅依靠共享模型对各种类型的行动进行质量评估是具有挑战性的。为了解决这些问题,我们提出了一种新的AQA网络结构,这是第一个通过分类模型集成多模型功能的网络结构。具体来说,我们利用预先训练的I3D模型配备了一个自注意块进行分类。这允许我们仅使用一个模型来评估各种类型的操作。此外,我们在传统卷积神经网络中引入了自注意机制和多头注意机制。通过系统地替换传统卷积网络的最后几层,我们的模型获得了更大的能力来感知不同动作的全局协调。我们已经在AQA-7数据集上验证了我们方法的有效性。与其他流行的模型相比,我们的模型在保持较低的计算成本的同时取得了令人满意的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
10.70
自引率
0.00%
发文量
27
审稿时长
12 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信