{"title":"Exploring the Gradient for Video Quality Assessment","authors":"Hossein Motamednia, Pooryaa Cheraaqee, Azadeh Mansouri","doi":"10.1109/MVIP49855.2020.9116869","DOIUrl":null,"url":null,"abstract":"This paper presents an algorithm which incorporates spatial and temporal gradients for full reference video quality assessment. In the proposed method the frame-based gradient magnitude similarity deviation is calculated to form the spatial quality vector. To capture the temporal distortion, the similarity of frame difference is measured. In the proposed method, we extract the worst scores in both the spatial and temporal vectors by introducing the variable-length temporal window for max-pooling operation. The resultant vectors are then combined to form the final score. The performance of the proposed method is evaluated on LIVE SD and EPFL- PoliMI datasets. The results clearly illustrate that, despite the computational efficiency, the predictions are highly correlated with human visual system.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 International Conference on Machine Vision and Image Processing (MVIP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MVIP49855.2020.9116869","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
This paper presents an algorithm which incorporates spatial and temporal gradients for full reference video quality assessment. In the proposed method the frame-based gradient magnitude similarity deviation is calculated to form the spatial quality vector. To capture the temporal distortion, the similarity of frame difference is measured. In the proposed method, we extract the worst scores in both the spatial and temporal vectors by introducing the variable-length temporal window for max-pooling operation. The resultant vectors are then combined to form the final score. The performance of the proposed method is evaluated on LIVE SD and EPFL- PoliMI datasets. The results clearly illustrate that, despite the computational efficiency, the predictions are highly correlated with human visual system.