{"title":"Video abstraction inspired by human visual attention models","authors":"S. O. Gilani, Mohsin Jamil","doi":"10.1109/IEMCON.2018.8614889","DOIUrl":null,"url":null,"abstract":"This paper describes an application of three different state-of-the-art human inspired visual attention models to video abstraction. Two types of video abstractions, video skim and key frame extraction, are performed over three different genres of videos. Qualitative and quantitative results are reported based on user studies and statistical tests. A comparison is made with human made video abstraction set to benchmark the current analysis. We report some abstraction similarities at scene level showing that all three models are successful in capturing semantic content despite having anatomical differences. However different models are more suitable for different genres of videos.","PeriodicalId":368939,"journal":{"name":"2018 IEEE 9th Annual Information Technology, Electronics and Mobile Communication Conference (IEMCON)","volume":"50 1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE 9th Annual Information Technology, Electronics and Mobile Communication Conference (IEMCON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IEMCON.2018.8614889","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper describes an application of three different state-of-the-art human inspired visual attention models to video abstraction. Two types of video abstractions, video skim and key frame extraction, are performed over three different genres of videos. Qualitative and quantitative results are reported based on user studies and statistical tests. A comparison is made with human made video abstraction set to benchmark the current analysis. We report some abstraction similarities at scene level showing that all three models are successful in capturing semantic content despite having anatomical differences. However different models are more suitable for different genres of videos.