Y. Dhamecha, S. Gadekara, S. Deshmukh, Y. Haribhakta
{"title":"基于特征向量聚类的视频摘要","authors":"Y. Dhamecha, S. Gadekara, S. Deshmukh, Y. Haribhakta","doi":"10.2139/ssrn.3734732","DOIUrl":null,"url":null,"abstract":"With ever-growing utilization of online and offline videos and increasing video content, Video Summarization serves as the best aid for video browsing. It involves domain explicit semantic comprehension of a video and understanding of user expectations. Generally, video summarization systems include extracting video features, analyzing the visual variations and selecting video frames. Over the years, various methodologies have been developed for the same. Different supervised and unsupervised algorithms have been established and these models have been trained on various factors or various rewards. The challenges these methods face stand as a motivation for the approach this paper discusses. Like in many cases, summary frames may be repeated if some scene or concept appears more than once. This paper presents a novel approach based on clustering of video frames based on their feature vectors. The clustering takes into consideration the semantic factor of video frames. Each concept cluster gives a representative frame which then forms the summary set, here concept cluster refers to the independent entity present in a video which can be easily distinguished by another concept or entity. This entity can be a scene of a mountain or different persons. It also aims to increase system performance by removing the redundancy. The system is developed using a CNN for feature extraction and a clustering algorithm that takes into consideration the similarity factor between these vectors. The model is evaluated on the measures Precision and Recall and tested on the VSUMM dataset. The results outperform some of the established methodologies and serve the summarization purpose.","PeriodicalId":11974,"journal":{"name":"EngRN: Engineering Design Process (Topic)","volume":"30 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2020-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Video Summarization Using Feature Vector Clustering\",\"authors\":\"Y. Dhamecha, S. Gadekara, S. Deshmukh, Y. Haribhakta\",\"doi\":\"10.2139/ssrn.3734732\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With ever-growing utilization of online and offline videos and increasing video content, Video Summarization serves as the best aid for video browsing. It involves domain explicit semantic comprehension of a video and understanding of user expectations. Generally, video summarization systems include extracting video features, analyzing the visual variations and selecting video frames. Over the years, various methodologies have been developed for the same. Different supervised and unsupervised algorithms have been established and these models have been trained on various factors or various rewards. The challenges these methods face stand as a motivation for the approach this paper discusses. Like in many cases, summary frames may be repeated if some scene or concept appears more than once. This paper presents a novel approach based on clustering of video frames based on their feature vectors. The clustering takes into consideration the semantic factor of video frames. Each concept cluster gives a representative frame which then forms the summary set, here concept cluster refers to the independent entity present in a video which can be easily distinguished by another concept or entity. This entity can be a scene of a mountain or different persons. It also aims to increase system performance by removing the redundancy. The system is developed using a CNN for feature extraction and a clustering algorithm that takes into consideration the similarity factor between these vectors. The model is evaluated on the measures Precision and Recall and tested on the VSUMM dataset. The results outperform some of the established methodologies and serve the summarization purpose.\",\"PeriodicalId\":11974,\"journal\":{\"name\":\"EngRN: Engineering Design Process (Topic)\",\"volume\":\"30 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-11-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"EngRN: Engineering Design Process (Topic)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.2139/ssrn.3734732\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"EngRN: Engineering Design Process (Topic)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2139/ssrn.3734732","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Video Summarization Using Feature Vector Clustering
With ever-growing utilization of online and offline videos and increasing video content, Video Summarization serves as the best aid for video browsing. It involves domain explicit semantic comprehension of a video and understanding of user expectations. Generally, video summarization systems include extracting video features, analyzing the visual variations and selecting video frames. Over the years, various methodologies have been developed for the same. Different supervised and unsupervised algorithms have been established and these models have been trained on various factors or various rewards. The challenges these methods face stand as a motivation for the approach this paper discusses. Like in many cases, summary frames may be repeated if some scene or concept appears more than once. This paper presents a novel approach based on clustering of video frames based on their feature vectors. The clustering takes into consideration the semantic factor of video frames. Each concept cluster gives a representative frame which then forms the summary set, here concept cluster refers to the independent entity present in a video which can be easily distinguished by another concept or entity. This entity can be a scene of a mountain or different persons. It also aims to increase system performance by removing the redundancy. The system is developed using a CNN for feature extraction and a clustering algorithm that takes into consideration the similarity factor between these vectors. The model is evaluated on the measures Precision and Recall and tested on the VSUMM dataset. The results outperform some of the established methodologies and serve the summarization purpose.