基于词袋模型的过去视频对象检索

Manh-Tien Nguyen-Hoang, Tu-Khiem Le, Van-Tu Ninh, Quoc-Huu Che, Vinh-Tiep Nguyen, M. Tran
{"title":"基于词袋模型的过去视频对象检索","authors":"Manh-Tien Nguyen-Hoang, Tu-Khiem Le, Van-Tu Ninh, Quoc-Huu Che, Vinh-Tiep Nguyen, M. Tran","doi":"10.1109/ICCAIS.2017.8217565","DOIUrl":null,"url":null,"abstract":"Together with the technology advancement, Computer Vision plays an important role in enhancing smart computing systems to help people overcome obstacles in their daily lives. One of the common troublesome problems is human memorization ability, especially memorizing things such as personal items. It is annoying for people to waste their time finding lost items manually by recall or notes. This motivates the authors to propose a solution that can help a user find an item that he or she already saw but vaguely remembers where and when it appeared in the past. The user simply provides our system a single image of that item, then the system retrieves a rank list of visual scenes that may contain the item from video recorded implicitly during user's daily activities. Our method is based on Bag-of-Words model, one of the most famous methods in image retrieval. We first conduct experiments to find the appropriate parameters and configurations of Bag-of-Words system for visual instance search. Then we perform experiments with 110 visual queries of 30 common objects in real video with 2837 shots recorded during daily activities of volunteers. Experimental results show that for all 30/30 categories of objects, our system can help users find their objects of interest just by looking into the top 10 video shots retrieved from recorded video with the balance accuracy from 50 to 80%. This demonstrates the potential use of our method to help people remind of their items in an easy and comfortable way.","PeriodicalId":410094,"journal":{"name":"2017 International Conference on Control, Automation and Information Sciences (ICCAIS)","volume":"53 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Object retrieval in past video using bag-of-words model\",\"authors\":\"Manh-Tien Nguyen-Hoang, Tu-Khiem Le, Van-Tu Ninh, Quoc-Huu Che, Vinh-Tiep Nguyen, M. Tran\",\"doi\":\"10.1109/ICCAIS.2017.8217565\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Together with the technology advancement, Computer Vision plays an important role in enhancing smart computing systems to help people overcome obstacles in their daily lives. One of the common troublesome problems is human memorization ability, especially memorizing things such as personal items. It is annoying for people to waste their time finding lost items manually by recall or notes. This motivates the authors to propose a solution that can help a user find an item that he or she already saw but vaguely remembers where and when it appeared in the past. The user simply provides our system a single image of that item, then the system retrieves a rank list of visual scenes that may contain the item from video recorded implicitly during user's daily activities. Our method is based on Bag-of-Words model, one of the most famous methods in image retrieval. We first conduct experiments to find the appropriate parameters and configurations of Bag-of-Words system for visual instance search. Then we perform experiments with 110 visual queries of 30 common objects in real video with 2837 shots recorded during daily activities of volunteers. Experimental results show that for all 30/30 categories of objects, our system can help users find their objects of interest just by looking into the top 10 video shots retrieved from recorded video with the balance accuracy from 50 to 80%. This demonstrates the potential use of our method to help people remind of their items in an easy and comfortable way.\",\"PeriodicalId\":410094,\"journal\":{\"name\":\"2017 International Conference on Control, Automation and Information Sciences (ICCAIS)\",\"volume\":\"53 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2017 International Conference on Control, Automation and Information Sciences (ICCAIS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCAIS.2017.8217565\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 International Conference on Control, Automation and Information Sciences (ICCAIS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCAIS.2017.8217565","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

随着科技的进步,计算机视觉在增强智能计算系统以帮助人们克服日常生活中的障碍方面发挥着重要作用。一个常见的麻烦问题是人类的记忆能力,特别是记忆诸如个人物品之类的东西。人们把时间浪费在手工寻找丢失的物品上,通过回忆或笔记是很烦人的。这促使作者提出了一种解决方案,可以帮助用户找到他或她已经看到的项目,但模糊地记得它在过去出现的时间和地点。用户只需向我们的系统提供该物品的单个图像,然后系统从用户日常活动期间隐式录制的视频中检索可能包含该物品的视觉场景的排序列表。我们的方法是基于词袋模型,这是图像检索中最著名的方法之一。我们首先通过实验找到了适合于视觉实例搜索的Bag-of-Words系统的参数和配置。然后,我们对真实视频中的30个常见物体进行了110次视觉查询,并记录了志愿者日常活动中的2837个镜头。实验结果表明,对于所有30/30类物体,我们的系统可以帮助用户通过查看从录制视频中检索到的前10个视频镜头来找到他们感兴趣的物体,平衡精度在50%到80%之间。这证明了我们的方法的潜在用途,帮助人们以一种简单舒适的方式提醒他们的物品。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Object retrieval in past video using bag-of-words model
Together with the technology advancement, Computer Vision plays an important role in enhancing smart computing systems to help people overcome obstacles in their daily lives. One of the common troublesome problems is human memorization ability, especially memorizing things such as personal items. It is annoying for people to waste their time finding lost items manually by recall or notes. This motivates the authors to propose a solution that can help a user find an item that he or she already saw but vaguely remembers where and when it appeared in the past. The user simply provides our system a single image of that item, then the system retrieves a rank list of visual scenes that may contain the item from video recorded implicitly during user's daily activities. Our method is based on Bag-of-Words model, one of the most famous methods in image retrieval. We first conduct experiments to find the appropriate parameters and configurations of Bag-of-Words system for visual instance search. Then we perform experiments with 110 visual queries of 30 common objects in real video with 2837 shots recorded during daily activities of volunteers. Experimental results show that for all 30/30 categories of objects, our system can help users find their objects of interest just by looking into the top 10 video shots retrieved from recorded video with the balance accuracy from 50 to 80%. This demonstrates the potential use of our method to help people remind of their items in an easy and comfortable way.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信