{"title":"DA-YOLOv5: Improved YOLOv5 based on Dual Attention for Object Detection on Coal Chemical Industry","authors":"Yan Wang, Haijiang Zhu, Yutong Liu","doi":"10.1109/PRMVIA58252.2023.00016","DOIUrl":"https://doi.org/10.1109/PRMVIA58252.2023.00016","url":null,"abstract":"The wearing inspection of personnel’s safety protective clothing has important practical significance in the safety production of coal chemical plants. Manual detection or traditional target detection methods are utilized in coal chemical plants for personnel’s safety detection at the moment. However, the clothing detection accuracy is seriously reduced due to the installation position of cameras and the change of light intensity in coal chemical plants. An dual attention based on YOLOv5 is proposed on coal chemical for object detection. Two attention modules, including Efficient Channel Attention (ECA) and Pyramid Split Attention (PSA) module, are integrated into the Spatial Pyramid Pooling (SPP) module and Bottleneck module of this YOLOv5 network. Thus, more global context information is obtained to make up for the lack of global convolution, and the ability to extract features and learn multi-scale information is enhanced. Safety helmet wearing detect data set (SHWD) and self-made data set in our work are utilized to display the improved method’s effectiveness. Compared with the original YOLOv5 algorithm, the improved method achieved an average accuracy increase of 2.7% at different thresholds. Numerous comparative experiments further verify the feasibility of the improved method.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126902769","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Garbage Classification and Detection Based on Improved YOLOv7 Network","authors":"Gengchen Yu, Birui Shao","doi":"10.1109/prmvia58252.2023.00024","DOIUrl":"https://doi.org/10.1109/prmvia58252.2023.00024","url":null,"abstract":"With the improvement of people’s living standards, garbage classification is gradually forced. However, due to people’s awareness and knowledge, the classification accuracy and disposal of garbage are difficult to keep pace with guideline changes. With the consideration of the problems of low efficiency, heavy task and poor environment of garbage manual classification, an improved YOLOv7 target detection method is proposed to realize the effective classification of garbage. In this study, the recursive gated convolutional gnconv was used to establish the HorNet network architecture, and the model was trained by making specific data sets. The C3HB module is added to the YOLO model, and the pooling layer is optimized to replace SPPFCSPC to improve the detection accuracy of the target. The experimental results show that the garbage detection and classification method proposed in this study has excellent accuracy. Experiments show that the map value, accuracy and recall rate of the proposed model on garbage datasets are 99.25%, 99.33% and 98.03%, respectively, which are 1.50%, 3.99% and 1.41% higher than those of YOLOv7. The overall results are better than the original model.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115657079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Tag-aware Recommendation Algorithm Based on Deep Learning and Multi-objective Optimization","authors":"Yi Zuo, Yun Zhou, Shengzong Liu, Yupeng Liu","doi":"10.1109/prmvia58252.2023.00013","DOIUrl":"https://doi.org/10.1109/prmvia58252.2023.00013","url":null,"abstract":"Social tagging information to describe characteristics. Recent systems introduce tagging user preferences and item work shows that the recommendation accuracy can be remarkably promoted when tag information is handled properly. However, other performance indicators of recommendations, such as diversity and novelty, are also of great importance in practice. Thus, we propose a two-stage tag-aware multi-objective framework for providing accurate and diversity recommendations. Specifically, we formulate a tag-based recommendation algorithm via deep learning to generate accurate items and abstract effective tag-based potential features for users and items. According to these features, two conflicting objectives are designed to estimate the recommendation accuracy and diversity, respectively. By optimizing these two objectives simultaneously, the designed multi-objective recommendation model can pro-vide a set of recommendation lists for each user. Comparative experiments verify that the proposed model is promising to generate improved recommendations in terms of accuracy and diversity.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130341459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Digital Protection and Virtual Display Technology of Ceramic Art","authors":"Linghao Cai","doi":"10.1109/prmvia58252.2023.00025","DOIUrl":"https://doi.org/10.1109/prmvia58252.2023.00025","url":null,"abstract":"Ceramic art has been passed down to the present day. It reflects the wisdom of ancient craftsmen and artists and is a treasure of Chinese culture. People with different cultural backgrounds and different artistic cultivation, their works are rooted in tradition and bold innovation. They are not only traditional and highly skilled skills, but also the essence of national culture. In recent years, with the rapid development of digital media technology, it has provided new development opportunities for the protection and inheritance of intangible cultural heritage. With the intervention of digital technology, the extension of ceramic design is also constantly extending, which requires ceramic designers to continuously expand their knowledge and combine multiple professional subject theories to enrich the connotation of their works. An excellent pottery work is not only a superficial artistic expression, but also a deep cultural heritage and innovative digital media performance. This article will discuss that in the digital age, the digital form of ceramics is obtained through three-dimensional scanning, and the ceramic art is digitally protected; and then the ceramic art is displayed through virtual technology and holographic imaging technology.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122810186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Modern Techniques for Rumor Detection from the Perspective of Natural Language Processing","authors":"Xinjia Xie, Shun Gai, Han Long","doi":"10.1109/PRMVIA58252.2023.00035","DOIUrl":"https://doi.org/10.1109/PRMVIA58252.2023.00035","url":null,"abstract":"Rumor detection on online social network (OSN) aims to help people retrieve reliable information and prevent public panic when emergencies occur suddenly. However, it is a waste of human efforts to detect rumors from the rapid growth of large-scale datasets. Due to the development of artificial intelligence, many architectures and frameworks are proposed to provide solutions for this issue. The first proposed traditional feature related methods are time-consuming and heavily depend on well-designed features, which calls for novel methods to detect rumors more efficiently. Thus deep neural networks related methods are successively born, and recent research on propagation related methods has captured much attention of both academia and industry. However, there lacks a systematic and global survey in the field of modern rumor detection. In this paper, we introduce rumors and OSN, and then present a comprehensive study of rumor detection methods on OSN, classifying them according to their search approaches and providing a comparison of the selected works. Finally, this survey deliver unique views on key challenges and several future research directions of rumor detection on OSN, such as multi-task learning, multi-modal detection and developing standard datasets and benchmarks. This work is supported by the Department of System Science, College of Liberal Arts and Sciences in National University of Defense Technology.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128990580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Knowledge Distillation by Multiple Student Instance Interaction","authors":"Tian Ni, Haoji Hu","doi":"10.1109/prmvia58252.2023.00038","DOIUrl":"https://doi.org/10.1109/prmvia58252.2023.00038","url":null,"abstract":"Knowledge distillation is an efficient method in neural network compression, which transfers the knowledge from a high-capacity teacher network to a low-capacity student network. Previous approaches follow the ‘one teacher and one student’ paradigm, which neglects the possibility that interaction of multiple students could boost the distillation performance. In this paper, we propose a novel approach by simultaneously training multiple instances of a student model. By adding the similarity and diversity losses into the baseline knowledge distillation and adaptively adjusting the proportion of these losses according to accuracy changes of multiple student instances, we build a distillation system to make students collaborate and compete with each other, which improves system robustness and performance. Experiments show superior performance of the proposed method over existing offline and online distillation schemes on datasets with various scales.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126419042","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Comparative Study Of CART Algorithm For Forecasting","authors":"Juanqin Yan, Quan Zhou, Ya Xiao, Bin Pan","doi":"10.1109/PRMVIA58252.2023.00028","DOIUrl":"https://doi.org/10.1109/PRMVIA58252.2023.00028","url":null,"abstract":"CART algorithm is a tree structure used for classification rules in the form of decision tree from a group of unordered and irregular cases. Compared with other classification methods, it has the advantage that a busy large amount of data can is classified yen fully, and then valuable potential information can be found. The method is simple and intuitive, with fast classification speed and high accuracy, which is suitable for large-scale data processing. Moreover, the algorithm process is easy to understand and can though express the importance of attributes praying attributes. The significant sensitivity and unpredictability of house price make it difficult to construct its forecasting model. In this paper, through an example of house price, the influencing factors of house price are deeply analyzed and the existing research results are systematically sorted out, and the decision tree CART detailed is used to build a molybdenum metal price algorithm model and forecast the actual price. By comparing and analyzing the results by using Not principles, the average absolute error is 4.03%, and the accuracy rate of foreforetrend forecasting trend can reach 94.8%, which shows that the algorithm is not only not intuitive and intuitive, but also reasonable and reliable.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125737698","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Attention-Based Recursive Autoencoder For Sentence-Level Sentiment Classification","authors":"Jiayi Sun, Mingbo Zhao","doi":"10.1109/PRMVIA58252.2023.00050","DOIUrl":"https://doi.org/10.1109/PRMVIA58252.2023.00050","url":null,"abstract":"Sentiment analysis is a crucial task in the research of natural language handling. Traditional machine learning approaches frequently employ bag-of-word representations that do not capture complex linguistic phenomena. The recursive autoencoder (RAE) method can availably learn the vector space representation of phrases, which is superior to other sentiment prediction methods on commonly used data sets. However, during the learning process, extensive label data is often required to label each node. In addition, RAE uses greedy strategies to merge adjacent words, it is difficult to capture long-distance and deeper semantic information. We put forward a semi-supervised approach that combines the SenticNet lexicon to train the recursive autoencoder for calculating the sentiment orientation of each node, and incorporates an attention mechanism to capture the contextual relationship between the words in a sentence. Experiments prove that the model proposed in this paper outperforms RAE and other models.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125896713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Surface deformation monitoring based on DINSAR technique","authors":"Xia Yu, YU Peng, Le Xia, Yuanrong He","doi":"10.1109/prmvia58252.2023.00056","DOIUrl":"https://doi.org/10.1109/prmvia58252.2023.00056","url":null,"abstract":"In this paper, we monitor the surface deformation of Helan Mountains by using the DInSAR (Differential Interferometric Synthetic Aperture Radar) technology and Sentinel-1 SAR data from December 2019 to December 2021. The surface deformation of the Helan Mountain National Natural Reserve with a study area extending to 1935 km2 are observed. The findings indicate that the surface of Helan Mountain Reserve is rising in the east and sinking in the west, with no obvious increasing tendency in the north or south of Helan Mountain. Additionally, the vertical deformation map created by D-InSAR processing is used to monitor two monitoring cycles with significant deformations in June 2020 and December 2021. Furthermore, Helan Mountain has experienced two earthquakes with magnitudes of 3 or greater, according to the differential interference technique. An important decision-making basis for disaster prevention and mitigation can be provided by the deformation data of the ground surface obtained by the InSAR technology.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114063899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Spatio-Temporal-based Context Fusion for Video Anomaly Detection","authors":"Chao Hu, Weibin Qiu, Weijie Wu, Liqiang Zhu","doi":"10.1109/PRMVIA58252.2023.00037","DOIUrl":"https://doi.org/10.1109/PRMVIA58252.2023.00037","url":null,"abstract":"Video anomaly detection (VAD) detects target objects such as people and vehicles to discover abnormal events in videos. There are abundant spatio-temporal context information in different objects of videos. Most existing methods pay more attention to temporal context than spatial context in VAD. The spatial context information represents the relationship between the detection target and surrounding targets. Anomaly detection makes a lot of sense. To this end, a video anomaly detection algorithm based on target spatio-temporal context fusion is proposed. Firstly, the target in the video frame is extracted through the target detection network to reduce background interference. Then the optical flow map of two adjacent frames is calculated. Motion features are used multiple targets in the video frame to construct spatial context simultaneously, re-encoding the target appearance and motion features, and finally reconstructing the above features through the spatiotemporal dual-stream network, and using the reconstruction error to represent the abnormal score. The algorithm achieves frame-level AUCs of 98.5% on UCSDped2 and 86.3% on Avenue datasets. On UCSDped2 dataset, the spatio-temporal dual-stream network improves frames by 5.1% and 0.3%, respectively, compared to the temporal and spatial stream networks. After using spatial context encoding, the frame-level AUC is enhanced by 1%, which verifies the method’s effectiveness.","PeriodicalId":221346,"journal":{"name":"2023 International Conference on Pattern Recognition, Machine Vision and Intelligent Algorithms (PRMVIA)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-05-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126900951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}