P. Nguyen, Tien Do, Anh-Thu Nguyen-Thi, T. Ngo, Duy-Dinh Le, T. Nguyen
{"title":"Clustering web video search results with convolutional neural networks","authors":"P. Nguyen, Tien Do, Anh-Thu Nguyen-Thi, T. Ngo, Duy-Dinh Le, T. Nguyen","doi":"10.1109/NICS.2016.7725638","DOIUrl":null,"url":null,"abstract":"Convolutional Neural Networks (CNNs) have been established as a powerful class of models for image recognition problems giving state-of-the-art results on recognition, detection, segmentation, classification and retrieval. Encouraged by these results, we develop our previous work [14] by implementing deep neural network architecture for extracting and representing visual features to improve the clustering quality of web video search results. Experiments were conducted on a dataset published in [14]. This dataset includes 1580 videos from 18 queries issued to the YouTube search engine. Our method exhibits significant performance improvements over the previously published result evaluated by Entropy measure (23.27% vs. 39.46%) and Purity measure (77.09% vs. 61.50%).","PeriodicalId":347057,"journal":{"name":"2016 3rd National Foundation for Science and Technology Development Conference on Information and Computer Science (NICS)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 3rd National Foundation for Science and Technology Development Conference on Information and Computer Science (NICS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NICS.2016.7725638","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
Convolutional Neural Networks (CNNs) have been established as a powerful class of models for image recognition problems giving state-of-the-art results on recognition, detection, segmentation, classification and retrieval. Encouraged by these results, we develop our previous work [14] by implementing deep neural network architecture for extracting and representing visual features to improve the clustering quality of web video search results. Experiments were conducted on a dataset published in [14]. This dataset includes 1580 videos from 18 queries issued to the YouTube search engine. Our method exhibits significant performance improvements over the previously published result evaluated by Entropy measure (23.27% vs. 39.46%) and Purity measure (77.09% vs. 61.50%).
卷积神经网络(cnn)已经被建立为一类强大的图像识别问题模型,在识别、检测、分割、分类和检索方面提供了最先进的结果。受这些结果的鼓舞,我们通过实现用于提取和表示视觉特征的深度神经网络架构来改进web视频搜索结果的聚类质量,从而发展了我们之前的工作[14]。实验是在[14]发表的数据集上进行的。该数据集包括来自YouTube搜索引擎的18个查询的1580个视频。与先前发表的熵测度(23.27% vs. 39.46%)和纯度测度(77.09% vs. 61.50%)评价结果相比,我们的方法表现出显著的性能改进。