{"title":"A Display Method of Large Underwater Photo-Mosaics Based on Pyramid Model of Tiled Images","authors":"Nannan Liu, Xiaoming Li","doi":"10.1109/CCAI50917.2021.9447465","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447465","url":null,"abstract":"In deep seafloor exploration, optical imaging provides short-range, high resolution visual information. In order to obtain a wide range of detailed visual information, it is common to stitch multiple images into a photo-mosaic which could reach tens of billions of pixels in size. Due to hardware and software constraints, it is very difficult even impossible to browse and display such a large image with existing image viewers. In this paper, we propose a display method based on pyramid model and develop a super large image display system of tiled images dedicated to underwater photo-mosaics display. Before the image can be displayed, the first step is to construct and store the multi-resolution hierarchical model of tiles, in which sub-mosaics of the same size are original tiles with the highest resolution. Similar to quadtree coding, each tile is encoded based on its location in the pyramid. Then all tile data are stored in MongoDB database. Each record in MongoDB is a key-value pair structure corresponding to a specific tile, in which key is the encoding of a tile and value is the tiled image data stored in binary stream format. Image display is implemented based on graphical user interface. By mouse operation, the system can display images at different resolutions and browse different part of the image. Based on the pyramid model and key-value storage structure, there are more than 60,000 high-definition tiled images in our application. The full panorama resolution is 16.2 billion pixels, about 45GB in RAM, but only the tiles displayed in the current window need to be loaded. Hence, our system reduces the memory requirement greatly and makes image browsing more smoothly.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115705019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Correlation Filter Tracking Algorithm Incorporating Time Information","authors":"Xiaoshuo Jia, Shangyou Zeng","doi":"10.1109/CCAI50917.2021.9447479","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447479","url":null,"abstract":"A target tracking algorithm that incorporates time information into the relevant filter is proposed, LTCF. LTCF is a combination of LT model and CFM model. We first analyze from the tag data that the target has time continuous characteristics and relative time information during the movement. Then, the LT model is designed on the basis of LSTM algorithm, which mainly to predict the time information about the target of the next frame. Finally, the information which predicted form LT model is combined with the CFM model, which improves the calculation time of the filter on the one hand, and reduces the overall algorithm parameters on the other. Experiment results show that compared with STRCF and SRCF, the accuracy of the LTCF algorithm is improved by 7% and 11% respectively, and LTCF algorithm model also reduces the model size while ensuring real-time tracking.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116700977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Simple-NMS: Improved Pedestrian Detection with New Constraints","authors":"Li Tian, Zhaogong Zhang","doi":"10.1109/CCAI50917.2021.9447459","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447459","url":null,"abstract":"Non-maximum suppression (NMS) is an indispensable post-processing step in object detection. After a picture is detected by the target, a large number of redundant anchors will be obtained, so we need to go through NMS to process these repeated anchors. Its essence is to extract the target detection frame with high confidence and suppress the false detection frame with low confidence. The biggest problem in the NMS algorithm is that it completely removes adjacent low-confidence detection frames. In some images with dense targets, this is very likely to cause missed detection and false detection. Therefore, we proposed the Simple-NMS algorithm, which adds two new thresholds to the original NMS, and sets a constraint condition different from the original based on the new threshold, and does not change the complexity of the traditional NMS algorithm, which can be very good Improve the effectiveness of NMS missed targets. The new NMS algorithm is improved on the standard data sets PASAL VOC2007 and MS COCO. In addition, the algorithm is simple and efficient, and can be easily integrated into any other object detection process.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"182 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114139502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Satisfactory Detection Framework on Online-shopping Comments","authors":"Hanqing Liu","doi":"10.1109/CCAI50917.2021.9447490","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447490","url":null,"abstract":"Online shopping is becoming one of the main channels of daily shopping. After buying, shoppers usually express their experience in the shopping process through ratings and written comments, which are also important bases for other shoppers to make choices between similar goods and similar sellers. The purpose of this paper is to design a structure for analyzing the text to quantify the consumer satisfaction hidden behind the review text, so as to guide sellers and consumers to a more refined understanding of the potential consumption behavior and make the comparison easier and more direct.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117014858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Collaborative Cyberinfrastructure for Research and Education Using Geospatial Model","authors":"Wei Wan, Xinwei Zhao, Xingqiang Du, Zhenkun Yang","doi":"10.1109/CCAI50917.2021.9447513","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447513","url":null,"abstract":"Many geospatial modelers around the world use remote sensing data to simulate environmental processes, ecological system and testing agricultural management scenarios. Once these tasks are complete including publication of results, the models generally are not published or made available to the public for further use and improvement. Sharing of models may open doors for new collaborations, and avoids duplication of efforts. For researchers, who are interested in sharing models, there are limited avenues to publishing their models to the wider community. Towards filling this gap, a prototype cyberinfrastructure is developed for publishing, sharing and running geospatial models in an interactive GIS-enabled web environment. Users can utilize the Hub to publish or upload their own models, search and download existing models or data developed by others, run simulations including processing, analysis and calibration using computing and storage resources provided by Cloud. This should open the way for easy development of a variety of web-enabled tools for probing and presenting digital geospatial data in ways that can help addressing issues around the globe.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129688169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Exploration on Channel-interactive Features in Silent Speech Recognition","authors":"Ming Zhang, You Wang, Guang Li, Xiang Sui","doi":"10.1109/CCAI50917.2021.9447446","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447446","url":null,"abstract":"In recent years, silent speech recognition (SSR) based on surface electromyography (sEMG) has been proven practical and achieved successes in previous researches. The channel correlations hidden in sEMG data may work in SSR, which is rarely researched. This paper proposed to apply frame-based local binary pattern (LBP) in feature extraction of multichannel sEMG. Data are preprocessed after collection by human subjects to obtain clean signals. Then time domain (TD) features and BLP are both extracted as two feature sets, which are reduced in dimension by the method of Linear Discriminant Analysis (LDA). Random forest is used as the classifier to provide the recognitions. Experimental results show that TD obtain better result than LBP, which is exceeded by using TD and LBP simultaneously with a correct rate of 0.85. This proves LBP contains unique features in sEMG, indicating interactions in channels are promising in SSR.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124643896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Research on Video Fire Detection Algorithm Based on Attention Mechanism","authors":"Wenting Ouyang, Y. Fang, Aimin Xiong, Haofei He","doi":"10.1109/CCAI50917.2021.9447460","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447460","url":null,"abstract":"To solve the problem that the current fire detection algorithm is susceptible to false positives and false negatives due to the influence of complex scenes and lighting conditions, this paper proposes a video fire recognition algorithm based on attention mechanism. Firstly, video stream datum are read frame by frame, then used for dynamic and static judgment. Fire features are extracted by static color segmentation and dynamic flicker judgment in the YCbCr color model. Secondly, multi-feature fusion and attention mechanism are introduced into the ResNet-50 network, which is used to automatically extract fire features to improve the detection effect of small targets. Finally, center point detection is used to predict fire and realize rapid and accurate fire positioning. The experimental results show that the processing speed of the algorithm is 68FPS and mAP is 89.1% on the fire dataset established in this paper. It has good robustness and meets the needs of real-time video fire detection.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126920882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Considering Filter Importance and Irreplaceability In Filter Pruning","authors":"Wei Lu","doi":"10.1109/CCAI50917.2021.9447516","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447516","url":null,"abstract":"Deep convolutional neural network (CNNs) have gained a great success in computer vision tasks. However, the computation and parameter storage costs of CNNs are very large, thus a large number of studies have tried to reduce the computation and parameters of CNNs. Quantization and pruning are the usual strategies for model compression. Previous filter pruning works usually prune filters with small norm because the common viewpoint is that the smaller norm of filters, the less contribution of filters. However, above strategy ignores that features extracted by those filters with large norm may be redundancy and the features extracted by small norm filters are irreplaceable. In this paper, in light of above findings, we propose a novel filter pruning method called Importance Diversity Filter Pruning (IDFP), which considers both filter contribution and filter irreplaceability. We conduct extensive experiments on CIFAR-10 and CIFAR-100 datasets. The results illustrate the effectiveness of our method.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127833640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jing-ying Zhao, Hai Guo, Yifan Liu, N. Dong, Xinyue Wang
{"title":"Tai Le Character Recognition Using Deep Convolution Neural Networks","authors":"Jing-ying Zhao, Hai Guo, Yifan Liu, N. Dong, Xinyue Wang","doi":"10.1109/CCAI50917.2021.9447467","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447467","url":null,"abstract":"Tai Le characters are used widely in Southwest China and Southeast Asia, and it is necessary to recognize Tai Le characters. In this paper, a Tai Le characters recognition model base on the deep convolution neural network (DCNN) has been established. In addition, to recognize Tai Le characters, the printed Tai Le database YDH2019.1 has been constructed, and image digital processing technology has been used to preprocess the sample images of YDH2019.1. In the modeling of DCNN, 50% data of YDH2019.1 has been used as train data, and the performance of the recognition, model has been evaluated on the remaining 50%. The results show that the recognition accuracy of Tai Le characters reaches 99.98%, which proved the effectiveness of DCNN as a printed Tai Le character recognition model.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"348 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131396901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xueli Wang, Xingtao Zhuang, Wei Zhang, Yunfang Chen, Yanchao Li
{"title":"Lightweight Real-time Object Detection Model for UAV Platform","authors":"Xueli Wang, Xingtao Zhuang, Wei Zhang, Yunfang Chen, Yanchao Li","doi":"10.1109/CCAI50917.2021.9447518","DOIUrl":"https://doi.org/10.1109/CCAI50917.2021.9447518","url":null,"abstract":"Real-time detecting objects on captured images on UAV (Unmanned Aerial Vehicle) platforms, rather than barely transmitting images back to supporting equipment for post-processing, is a core requirement for advanced UAV applications. However, due to limited computing capacity and memory of UAV platforms, it is very challenging to deploy real-time detection models on them. In addition, there are more small objects in aerial images, which makes it more difficult to detect accurately. To solve these problems, this paper brings dense connection to Yolo(You Only Look Once)v3 network, and proposes Yolo-LiteDense model. The backbone of Yolo-LiteDense is densely connected, which improves the performance of feature extraction. Then, we enforce channel pruning to Yolo-LiteDense model by pruning less informative channels with less scaling factors. After pruning, parameters and weight size of the model are compressed significantly, and inference time is also shortened. Evaluation results on VisDrone2018-DET show that parameters and weight size of Yolo-LiteDense are 83% and inference time is 30% less than Yolov3-SPP with comparable average precision. In addition, this paper also proposes the lighter version of Yolo-LiteDense, Yolo-DenseNano. Parameters and weight size of Yolo-LiteDense are 70% less than Yolov3-tiny with 2.68 times greater average precision.","PeriodicalId":121785,"journal":{"name":"2021 International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127662283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}