{"title":"Using Spatial-Temporal Attention for Video Quality Evaluation","authors":"Biwei Chi, Ruifang Su, Xinhui Chen","doi":"10.1155/2024/5514627","DOIUrl":null,"url":null,"abstract":"<div>\n <p>With the rapid development of media, the role of video quality assessment (VQA) is becoming increasingly significant. VQA has applications in many domains. For example, in the field of remote medical diagnosis, it can enhance the quality of video communication between doctors and patients. Besides, in sports broadcasting, it can improve video clarity. Within VQA, the human visual system (HVS) is a crucial component that should be taken into consideration. Considering that attention is guided by goal-driven and top-down factors, such as anticipated locations or some attractive frames within the video, we propose a blind VQA algorithm based on spatial-temporal attention model. Specifically, we first use two pretrained convolutional networks to extract low-level static-dynamic fusion features. Then, a spatial attention-guided model is established to get more representative features of frame-level quality perception. Next, through a temporal attention-guided model, the video-level features are obtained. Finally, the features are fed into a regression model to calculate the final video quality score. The experiments conducted on seven VQA databases reach the state-of-the-art performance, demonstrating the effectiveness of our proposed method.</p>\n </div>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":5.0000,"publicationDate":"2024-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1155/2024/5514627","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1155/2024/5514627","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
With the rapid development of media, the role of video quality assessment (VQA) is becoming increasingly significant. VQA has applications in many domains. For example, in the field of remote medical diagnosis, it can enhance the quality of video communication between doctors and patients. Besides, in sports broadcasting, it can improve video clarity. Within VQA, the human visual system (HVS) is a crucial component that should be taken into consideration. Considering that attention is guided by goal-driven and top-down factors, such as anticipated locations or some attractive frames within the video, we propose a blind VQA algorithm based on spatial-temporal attention model. Specifically, we first use two pretrained convolutional networks to extract low-level static-dynamic fusion features. Then, a spatial attention-guided model is established to get more representative features of frame-level quality perception. Next, through a temporal attention-guided model, the video-level features are obtained. Finally, the features are fed into a regression model to calculate the final video quality score. The experiments conducted on seven VQA databases reach the state-of-the-art performance, demonstrating the effectiveness of our proposed method.
期刊介绍:
The International Journal of Intelligent Systems serves as a forum for individuals interested in tapping into the vast theories based on intelligent systems construction. With its peer-reviewed format, the journal explores several fascinating editorials written by today''s experts in the field. Because new developments are being introduced each day, there''s much to be learned — examination, analysis creation, information retrieval, man–computer interactions, and more. The International Journal of Intelligent Systems uses charts and illustrations to demonstrate these ground-breaking issues, and encourages readers to share their thoughts and experiences.