{"title":"Gradient Domain Based Processing Method for Image Synthesis","authors":"Zhike Yi, Liang Liu, Jing Zhang, Shuai Li, A. Hao","doi":"10.1109/ICVRV.2018.00029","DOIUrl":"https://doi.org/10.1109/ICVRV.2018.00029","url":null,"abstract":"Digital image synthesis technology is an important technology in the field of computer graphics. It is widely used in various scenes such as digital image editing, film and television production, graphic design and so on. Gradient domain is an important concept in digital images. It can reflect the degree of variation of pixel values in each location of a digital image in its adjacent areas, and has a good perception effect on the edges of objects in digital images. Gradient-domain image synthesis method is an effective technique in the field of digital image synthesis. Its goal is to select a specific region from the source image and fuse it into the background image so that the images which before and after the fusion have no fusion traces. Vivid and natural. In this paper, the characteristics of digital image synthesis method based on gradient domain are analyzed. The Poisson clonal algorithm is used for image fusion. The characteristics of the core equations involved in Poisson clonal algorithm are studied. The equations are solved by combining the characteristics of coefficient matrix. The process is optimized and the solution of the linear equation is replaced by a Gaussian elimination method with a conjugate gradient method, which enables the digital image fusion process to be completed within a user-acceptable speed range.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"659 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113982140","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Virtual Reassembly of Fractured Bones for Orthopedic Surgery","authors":"Lei Wang, Junjun Pan, Qiangqiang Yao","doi":"10.1109/ICVRV.2018.00012","DOIUrl":"https://doi.org/10.1109/ICVRV.2018.00012","url":null,"abstract":"In orthopedic surgery, the traditional restoration of comminuted fracture basically relies on surgeon's manual work, which is usually intricate and error-prone. If fractured bones are poorly reassembled, the patient may suffer from more exposure to radiation and longer recovery time. More severely, some operation mistakes can cause sequela, such as joint dysfunction and infection. Therefore, orthopedic surgeons urgently need an intelligent assistance solution to improve the accuracy and reliability of fractured bones reassembly in procedure. This paper presents an automatic pipeline for virtual reassembly of fractured bones which are broken into pieces. It uses an intact bone as a template. We first reconstruct the 3D fractured bone from CT data using MIMICS. Then we analyze the shape structure of bone model through extracting key feature points and comparing descriptors. Finally, we search the correspondence between the fragments and template. The aligning is performed to ensure the fragments can join together. Compared with some semi-automatic reassembly methods for archaeological artifacts and forensic evidences, ours is fully automatic without human interactions and specialized to medical purposes. It can help orthopedic surgeons to make correct decisions in fractured bones reassembly.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127839109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Applicability Analysis on Three Interaction Paradigms in Immersive VR Environment","authors":"Zhuoran Li, Shiqi Zhang, M. Anwar, Junchang Wang","doi":"10.1109/ICVRV.2018.00024","DOIUrl":"https://doi.org/10.1109/ICVRV.2018.00024","url":null,"abstract":"The work intends to research the characteristics and applicability of three different types of interaction paradigms in immersive VR environment. For this purpose, a three-dimensional VR scene is designed based on the Unity3D simulation engine, and three types of interaction paradigms, including graphical user interface interaction paradigm, gesture interaction paradigm and voice interaction paradigm, are developed into this VR scene respectively. And the designed immersive and interactive scenes are experienced with HTC Vive. Taking into consideration the development of three interaction paradigms, the computational costs of the running scene with different type of interaction paradigm, and the user learning cost, the characteristics of the three types of interaction paradigms areexploredandcomparedwitheachotherforapplicabilityanalysis. Furthermore, the recommendation for selecting the appropriate interaction paradigm in different circumstances is provided according to analysis result.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115591613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Physically-Based Algorithm for Natural Rime Growth Simulation","authors":"Ye Li, Meng Yang, Gang Yang","doi":"10.1109/icvrv.2018.00021","DOIUrl":"https://doi.org/10.1109/icvrv.2018.00021","url":null,"abstract":"In order to show a common natural landscape in winter, the charm of rime, an algorithm for rime growth simulation based on a physical approach is proposed in this paper. This algorithm mainly simulates two main kinds of rimes' growth, especially their morphology and direction, under the condition of surrounding factors, including the air pressure, temperature and wind. Firstly, this algorithm calculates the length of rime by using the icing conductor model which is established according to fluid mechanics and thermodynamic principles, a fractal method is introduced into our algorithm to simulate the morphology of the crystalline rime, and a segmentation method is for the simulation of the needle-like rime's morphology; After that, wind field is simulated by adopting the function of Perlin Noise. Finally, it analyzes the offset details of the rime growth in the wind according to its material mechanics knowledge. The wind force is used to calculate the deviation of rime. Experimental results show us that our algorithm in this paper can simulate two kinds of rime realistically, effectively and efficiently.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"97 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133366267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Aircraft Detection Based on Multiple Scale Faster-RCNN","authors":"Wei Miao, Z. Luo","doi":"10.1109/ICVRV.2018.00026","DOIUrl":"https://doi.org/10.1109/ICVRV.2018.00026","url":null,"abstract":"Remote sensing image recognition has been widely used in civil and military fieldsDIn view of plenty of interference factors in remote-sensing aircraftCsuch as shadeCnoiseCthe changing of perspectiveCetc. An improved target recognition algorithm in remote sensing image based on Faster-RCNN is proposed which uses a standard Region Proposal Network (RPN) generation and incorporates feature maps from shallower convolution feature maps. Convolution neural network is adopted to recognize aircraft target in complex environment , enhance the global context and local information to avoid information loss in the process of feature extractionCwhich improves recognition rate. Simulation results show that the feasibility of aircraft target recognition algorithm in remoting sensing image and the scale and posture changes of target can be overcome.MeanwhileCthe proposed algorithm has higher recognition effect and stronger robustness than traditional Faster-RCNN and BP neural network and support vector machine ( SVM) methods.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"50 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130740408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Large-Scale Scene Display System Based on WebGL","authors":"W. Jiang, Yao Li, Yue Qi","doi":"10.1109/ICVRV.2018.00011","DOIUrl":"https://doi.org/10.1109/ICVRV.2018.00011","url":null,"abstract":"In recent decades, computer vision has been a hot area of computer research while 3D reconstruction is one of the hot topics. With the improvement of 3D reconstruction theory and the rapid development of computer hardware technology, the reconstructed 3D models are enlarging in scale and increasing in complexity. Combining the aerial image of UAV with the 3D reconstruction of sequence images, it is of great value to carry out relevant research on the problem of 3D reconstruction of outdoor scenes. The fast rendering of these large-scale scene models is a challenge. At the same time, with the advancement of the HMTL5 specification, browsers have become more and more powerful, and WebGL has gradually gained support from more and more devices. The 3D rendering of the BS architecture will be the future direction of the rendering engine. The browser does not need to install the client program, which is convenient for users to access and process in any environment. Therefore, the large-scale scene display system based on web architecture is more convenient. In this paper, we designed a real-time display system for large-scale scenes, and studied WebGL based rendering frame techniques.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128409468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Secure and Efficient Face-Recognition Scheme Based on Deep Neural Network and Homomorphic Encryption","authors":"Xiaodong Li, Qing Han, Xin Jin","doi":"10.1109/ICVRV.2018.00017","DOIUrl":"https://doi.org/10.1109/ICVRV.2018.00017","url":null,"abstract":"In recent years, with the maturity of face recognition technology, face recognition has been widely used in real life, raising concerns about the accuracy of face recognition results, the efficiency of face recognition and the safety of data. So we proposed a secure and efficient face-recognition scheme based on deep neural network and homomorphic encryption. The entire scheme is divided into two parts: the client and the server. The client obtains the face images. The server performs recognition. Face features are extracted using deep neural networks and then encrypted with the Paillier algorithm. The data of face features is transferred from the client to the server with encrypted mode and does not need to be decrypted in the entire recognition process. In the recognition process, we adopt a highly efficient secretive Hamming distance calculation method and introduce a parallel computing scheme to encrypt feature data and calculate the ciphertext Hamming distance, which greatly improves the recognition efficiency of the entire program. No messages are leaked between the client and the server on the entire scheme, which achieves the purpose of protecting privacy and security. Compared with the previous secure face recognition scheme, the experimental results show that we improve the accuracy of and the efficiency of recognition while ensuring security.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127926250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Robust Method for Hands Gesture Recognition from Egocentric Depth Sensor","authors":"Ye Bai, Yue Qi","doi":"10.1109/icvrv.2018.00015","DOIUrl":"https://doi.org/10.1109/icvrv.2018.00015","url":null,"abstract":"We present a method for robust and accurate hand pose recognition from egocentric depth cameras. Our method combines CNN based hand pose estimation and joint locations based hand gesture recognition. In pose estimation stage, we use a hand geometry prior network to estimate the hand pose. In gesture recognition stage, we defined a hand language which based on a set of pre-define basic propositions, obtained by applying four predicate types to the fingers and palm states. The hand language is used to convert the estimated joint location to hand gesture. Our experimental results indicate that the method enables robust and accurate gesture recognition in self-occlusion environment.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128016603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"NLQA Based Knowledge Search for Virtual Geographic Environment","authors":"B. Jiang, Liheng Tan, Xiaohui Chen, Wei Zhang","doi":"10.1109/icvrv.2018.00019","DOIUrl":"https://doi.org/10.1109/icvrv.2018.00019","url":null,"abstract":"Big data has injected new vitality into the virtual geographic environment (VGE). The intelligent virtual geographic environment system has put forward higher requirements for the interactivity and intelligent services. Firstly, in this paper, we have constructed a multi-level semantic conversion model, and enhanced the Chinese geographic knowledge graph by structured geographic information data such as the vector data of map. And then, based on Chinese geographic knowledge graph, we have proposed a bilateral LSTM-CRF model to achieve natural language question answering for VGE. Combing geographic knowledge base with virtual geographic scenes, we experimented the method. The experimental results prove that the method which is natural language question answering (NLQA) combined with the knowledge base can shorten the distance between people and virtual scenes and enhance the immersion and interactivity of VGE. It is an important way for virtual geographic environment to become intelligent.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"253 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116472875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Skeleton Capsule Net: An Efficient Network for Action Recognition","authors":"Yue Yu, Niehao Tian, Xiangru Chen, Ying Li","doi":"10.1109/ICVRV.2018.00022","DOIUrl":"https://doi.org/10.1109/ICVRV.2018.00022","url":null,"abstract":"Capsule network is a new type of deep learning method to improve the CNN module. Though it has performed quite well on classifying the MNIST dataset, there are few applications in other fields. Thus in this paper, we apply the capsule network on skeleton-based classification and propose a framework to explore the potential of it. Since the bottom layer of the capsule network is still based on convolution operation, we feed heatmap as well as raw skeleton data and reach good performance on convolution-based action recognition. Most researches take spatial and temporal features into consideration and they do help to recognition accuracy. We propose two different encapsulations to extract the spatial and temporal features of skeleton sequences. We perform our experiments on UT-Kinect and a portion of NTU RGB+D dataset, and we achieve best 87% accuracy on the NTU RGB+D dataset. We also find that the capsule network is suitable for the coarse-grained classification tasks. In a conclusion, not only the characteristics of capsule network are proved, but also an efficient method to recognize human action is realized.","PeriodicalId":159517,"journal":{"name":"2018 International Conference on Virtual Reality and Visualization (ICVRV)","volume":"192 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114987341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}