JudPriNet:基于语义关系和蒙特卡罗采样的视频转换检测

Bo Ma;Jinsong Wu;Wei Qi Yan
{"title":"JudPriNet:基于语义关系和蒙特卡罗采样的视频转换检测","authors":"Bo Ma;Jinsong Wu;Wei Qi Yan","doi":"10.23919/ICN.2024.0010","DOIUrl":null,"url":null,"abstract":"Video understanding and content boundary detection are vital stages in video recommendation. However, previous content boundary detection methods require collecting information, including location, cast, action, and audio, and if any of these elements are missing, the results may be adversely affected. To address this issue and effectively detect transitions in video content, in this paper, we introduce a video classification and boundary detection method named JudPriNet. The focus of this paper is on objects in videos along with their labels, enabling automatic scene detection in video clips and establishing semantic connections among local objects in the images. As a significant contribution, JudPriNet presents a framework that maps labels to “Continuous Bag of Visual Words Model” to cluster labels and generates new standardized labels as video-type tags. This facilitates automatic classification of video clips. Furthermore, JudPriNet employs Monte Carlo sampling method to classify video clips, the features of video clips as elements within the framework. This proposed method seamlessly integrates video and textual components without compromising training and inference speed. Through experimentation, we have demonstrated that JudPriNet, with its semantic connections, is able to effectively classify videos alongside textual content. Our results indicate that, compared with several other detection approaches, JudPriNet excels in high-level content detection without disrupting the integrity of the video content, outperforming existing methods.","PeriodicalId":100681,"journal":{"name":"Intelligent and Converged Networks","volume":"5 2","pages":"134-146"},"PeriodicalIF":0.0000,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10601664","citationCount":"0","resultStr":"{\"title\":\"JudPriNet: Video Transition Detection Based on Semantic Relationship and Monte Carlo Sampling\",\"authors\":\"Bo Ma;Jinsong Wu;Wei Qi Yan\",\"doi\":\"10.23919/ICN.2024.0010\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Video understanding and content boundary detection are vital stages in video recommendation. However, previous content boundary detection methods require collecting information, including location, cast, action, and audio, and if any of these elements are missing, the results may be adversely affected. To address this issue and effectively detect transitions in video content, in this paper, we introduce a video classification and boundary detection method named JudPriNet. The focus of this paper is on objects in videos along with their labels, enabling automatic scene detection in video clips and establishing semantic connections among local objects in the images. As a significant contribution, JudPriNet presents a framework that maps labels to “Continuous Bag of Visual Words Model” to cluster labels and generates new standardized labels as video-type tags. This facilitates automatic classification of video clips. Furthermore, JudPriNet employs Monte Carlo sampling method to classify video clips, the features of video clips as elements within the framework. This proposed method seamlessly integrates video and textual components without compromising training and inference speed. Through experimentation, we have demonstrated that JudPriNet, with its semantic connections, is able to effectively classify videos alongside textual content. Our results indicate that, compared with several other detection approaches, JudPriNet excels in high-level content detection without disrupting the integrity of the video content, outperforming existing methods.\",\"PeriodicalId\":100681,\"journal\":{\"name\":\"Intelligent and Converged Networks\",\"volume\":\"5 2\",\"pages\":\"134-146\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10601664\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Intelligent and Converged Networks\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10601664/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Intelligent and Converged Networks","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10601664/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

视频理解和内容边界检测是视频推荐的重要阶段。然而,以往的内容边界检测方法需要收集包括位置、演员、动作和音频在内的信息,如果缺少其中任何一个元素,都可能对结果产生不利影响。为了解决这一问题并有效检测视频内容中的过渡,本文介绍了一种名为 JudPriNet 的视频分类和边界检测方法。本文的重点是视频中的对象及其标签,从而实现视频片段中的场景自动检测,并建立图像中局部对象之间的语义联系。作为一项重大贡献,JudPriNet 提出了一个框架,可将标签映射到 "连续视觉词袋模型",从而对标签进行聚类,并生成新的标准化标签作为视频类型标签。这有助于视频片段的自动分类。此外,JudPriNet 还采用蒙特卡洛抽样方法对视频片段进行分类,并将视频片段的特征作为框架内的元素。这种方法将视频和文本内容无缝整合在一起,同时不影响训练和推理速度。通过实验,我们证明了具有语义连接的 JudPriNet 能够有效地对视频和文本内容进行分类。我们的结果表明,与其他几种检测方法相比,JudPriNet 在高级内容检测方面表现出色,不会破坏视频内容的完整性,优于现有方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
JudPriNet: Video Transition Detection Based on Semantic Relationship and Monte Carlo Sampling
Video understanding and content boundary detection are vital stages in video recommendation. However, previous content boundary detection methods require collecting information, including location, cast, action, and audio, and if any of these elements are missing, the results may be adversely affected. To address this issue and effectively detect transitions in video content, in this paper, we introduce a video classification and boundary detection method named JudPriNet. The focus of this paper is on objects in videos along with their labels, enabling automatic scene detection in video clips and establishing semantic connections among local objects in the images. As a significant contribution, JudPriNet presents a framework that maps labels to “Continuous Bag of Visual Words Model” to cluster labels and generates new standardized labels as video-type tags. This facilitates automatic classification of video clips. Furthermore, JudPriNet employs Monte Carlo sampling method to classify video clips, the features of video clips as elements within the framework. This proposed method seamlessly integrates video and textual components without compromising training and inference speed. Through experimentation, we have demonstrated that JudPriNet, with its semantic connections, is able to effectively classify videos alongside textual content. Our results indicate that, compared with several other detection approaches, JudPriNet excels in high-level content detection without disrupting the integrity of the video content, outperforming existing methods.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信