{"title":"Early Diagnosis of Alzheimer's Disease Using Deep Learning","authors":"Huanhuan Ji, Zhenbing Liu, W. Yan, R. Klette","doi":"10.1145/3341016.3341024","DOIUrl":"https://doi.org/10.1145/3341016.3341024","url":null,"abstract":"Alzheimer's disease (AD) leads to memory loss and impairment, which may cause further symptoms. It affects lives of patients seriously and is not curable, but early confirmation of AD may be helpful to start proper treatment so as to avoid further brain damage. Over the past decades, machine learning methods have been applied to the classification of AD with results based on manually prepared features and a classifier having a multiple-step architecture. Recently, with the development of deep learning, the end to- end process of neural networks has been employed for pattern classification. In this paper, we focus on early diagnosis of AD based on convolutional neural networks (ConvNets) by using magnetic resonance imaging (MRI). Image slices of gray matter and white matter from MRI have been used as the inputs for classification. Ensemble learning methods have been employed after the convolutional operations for improving the classification by combining outputs of deep learning classifiers [27]. Three base ConvNets were designed, implemented, and compared in this paper. Our method was evaluated based on a dataset from the Alzheimer's Disease Neuroimaging Initiative for the early diagnosis of this illness. In particular, the accuracy rates of our classifications have reached up to 97.65% for AD/mild cognitive impairment and 88.37% for mild cognitive impairment/normal control.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129011951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Deep Residual Network for Single Image Super-Resolution","authors":"Haimin Wang, K. Liao, B. Yan, Run Ye","doi":"10.1145/3341016.3341030","DOIUrl":"https://doi.org/10.1145/3341016.3341030","url":null,"abstract":"This paper proposes a Deep Residual Network for Single Image Super-Resolution (DRSR). We build a deep model using residual units that remove unnecessary modules. We can build deeper network at the same computing resources with the modified residual units. Experiments shows that deepening the network structure can fully utilize the image contextual information to improve the image reconstruction quality. The network learns both global residuals and local residuals, making the network easier to train. Our network directly extracts features from Low-Resolution (LR) images to reconstruct High-Resolution (HR) images. Computational complexity of the network is dramatically reduced in this way. Experiments shows that our network not only performs well in subjective visual effect but also achieves a high level in objective evaluation index.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121319189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Using Machine Vision to Command a 6-axis Robot Arm to Act on a Randomly Placed Zinc Die Cast Product","authors":"Luke Butters, Zezhong Xu, R. Klette","doi":"10.1145/3341016.3341018","DOIUrl":"https://doi.org/10.1145/3341016.3341018","url":null,"abstract":"The paper presents a method, documenting the machine vision techniques required, to automate a manual process at a local New Zealand manufacturing company. The system was required to monitor a conveyor belt for zinc die cast outputs. A robot arm was informed how to pick up the object. Die cast outputs appear face up or down and can be rotated 360°. Four experiments were conducted to determine how accurately the proposed system could detect if the cast was lying face up or down, determine the robot picking location, and determine the angle of direction while performing an error checking function. Circle Hough transform and principal component analysis, along with other object analysis techniques, were used on 2D datasets. Point cloud data was analysed to determine if the cast was face up or down when taken from a time-of-flight camera.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"386 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126735663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Design Therapy for Post-Stroke Patients with Robotics Tools and Principles of Mirror Neurons Using qEEG Parameter Analysis","authors":"R. Setiawan, Odilia Valentine, H. Zakaria","doi":"10.1145/3341016.3341027","DOIUrl":"https://doi.org/10.1145/3341016.3341027","url":null,"abstract":"Stroke causes neurological disorders such as reduced muscle motor skills, as well as cognitive, visual, and coordination functions, significantly. The reduced level of independence and mobility of a person can affect their quality of life. So, with rehabilitation program, it is expected that motor skills and cognitive function of stroke patients can be restored. This study focused on designing an integrated therapeutic device by using stationary cycles with the principle of mirror neurons and equipped with qEEG signal analysis for parameters comparison before and after therapy. Channels EEG used are F3, F4, C3, Cz, C4, P3, Pz, and P4 according to the rules of localization 10-20 Ref. The device also designed as a robotic system controlled by the movement intentions detected from EEG signals before the actual movement occurs. The Event-Related Potential (ERP) phenomenon mu waves were used to recognize the movement intentions. The device has been tested on normal subjects. To train the movement intention algorithm, the subject were asked to cycle and stop at prescribed time by audio command. qEEG parameters were displayed through a monitor for realtime observation by physical therapist. This research is part of the design process of the stroke therapy system which will then be tested in several stroke patients. Therapy programs for post-stroke patients using this device are expected to increase mobility of the affected limb by stroke.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115435715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Video Object Segmentation with 3D Convolution Network","authors":"Huiyun Tang, Pin Tao, Rui Ma, Yuanchun Shi","doi":"10.1145/3341016.3341031","DOIUrl":"https://doi.org/10.1145/3341016.3341031","url":null,"abstract":"We explore a novel method to realize semi-supervised video object segmentation with special spatiotemporal feature extracting structure. Considering 3-dimension convolution network can convolute a volume of image sequence, it is a distinct way to get both spatial and temporal information. Our network is composed of three parts, the visual module, the motion module and the decoder module. The visual module learns object appearance feature from object in the first frame for network to detect specific object in following image sequences. The motion module aims to get spatiotemporal information of image sequences with 3-dimension convolution network, which learns diversities of object temporal appearance and location. The purpose of decoder module is to get foreground object mask from output of visual module and motion module with concatenation and upsampling structure. We evaluate our model on DAVIS segmentation dataset[15]. Our model doesn't need online training compared with most detection-based methods because of visual module. As a result, it takes 0.14 second per frame to get mask which is 71 times faster than the state-of-art method-OSVOS[2]. Our model also shows better performance than most methods proposed in recent years and its meanIOU accuracy is comparable with state-of-art methods.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128561012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Alaa E. Abdel-Hakim, M. El-Melegy, Shreen K. Refaay
{"title":"Recognizing Faces in Shades of Gray","authors":"Alaa E. Abdel-Hakim, M. El-Melegy, Shreen K. Refaay","doi":"10.1145/3341016.3341029","DOIUrl":"https://doi.org/10.1145/3341016.3341029","url":null,"abstract":"Face recognition depends on relatively few distinguishing features, when compared with common facial features. This gives color information greater value to recognition and identification processes. However, dealing with grayscale facial images is a must in some cases, e.g. legacy images. In this paper, we investigate the effect of losing color information on face recognition. We propose a novel framework, which utilizes CNN-based colorization before a CNN classifier. The proposed framework is tested on LFW benchmark dataset. The evaluation results prove the success of the proposed framework in reducing the negative effect of dropping the color information on face recognition performance.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117015460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-focus Color Image Fusion using Laplacian Filter and Discrete Fourier Transformation with Qualitative Error Image Metrics","authors":"S. Khan, Muzammil Khan, Qiong Ran","doi":"10.1145/3341016.3341019","DOIUrl":"https://doi.org/10.1145/3341016.3341019","url":null,"abstract":"The Multi-focus image is limited Depth-of-Field (DOF) of an imaging system causes blur images when the sample is wider than the DOF of the optical system. Color multi-focus image fusion allows merging two multi-focus images and produces a composite image integrating complementary information to better understand the entire scene. This paper introduced a new two-level approach for color image fusion. In the first level, the multi-focus image is enhanced by Laplacian Filter (LF) technique with Discrete Fourier Transform (DFT). In the second level, the enhanced images are further processed by Stationary Wavelet Transform (SWT) and construct a more informative fused image. The experimental results on the color multi-focused images showed the capabilities and improved results of the proposed approach as compared with the traditional SWT method. The output image is evaluated using both a qualitative and the quantitative approach. The standard deviation, mean and entropy quantitative metrics are used to assess the performance of the proposed method. The article also introduced a new type of qualitative performance metric named as \"Qualitative Error Image (QEI)\" to evaluate the proposed method and assess future evaluation.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"333 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132335450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Y. Yamato, Hirofumi Noguchi, M. Kataoka, Takuma Isoda
{"title":"Proposal of Environment Adaptive Software","authors":"Y. Yamato, Hirofumi Noguchi, M. Kataoka, Takuma Isoda","doi":"10.1145/3341016.3341035","DOIUrl":"https://doi.org/10.1145/3341016.3341035","url":null,"abstract":"Recently, heterogeneous hardware such as GPU and FPGA is used in many systems and also IoT devices are increased repidly. However, to utilize heterogeneous hardware, the hurdles are currently high because of much technical skills. In order to break down such a situation, we think it is required in the future that application programmers only need to write logics to be processed, then software will adapt to the environments with heterogeneous hardware, to make it easy to utilize heterogeneous hardware and IoT devices. Therefore, in this paper, we propose environment adaptive software to operate an once written application with high performance by automatically converting the code and configuring setting so that we can utilize GPU, FPGA and IoT devices in the location to be deployed. We explain a processing flow and elemental technologies to achieve environment adaptive software. We also describe the details of elemental technologies such as automatic GPU offloading which are already under considered.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128064716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Chinese Rubbing Image Binarization based on Deep Learning for Image Denoising","authors":"Zhi-Kai Huang, Zhen-Ning Wang, Jun-Mei Xi, Ling-Ying Hou","doi":"10.1145/3341016.3341023","DOIUrl":"https://doi.org/10.1145/3341016.3341023","url":null,"abstract":"Aiming at the problem of the chinese rubbing image segmentation under a denoising algorithm based on deep convolutional neural network is proposed. Document enhancement and binarization is the main pre-processing step in document analysis process. At first, a feed-forward denoising convolutional neural networks as a pre-processing methods for document image has been used for denoise images of additive white Gaussian noise(AWGN). The residual learning mechanism is used to learn the mapping from the noisy image to the residual image between the noisy image and the clean image in the neural network training process. A median filtering has been employed for denoising 'salt and pepper' noise. Given the learned denoising and enhanced image, we compute the adaptive threshold image using local adaptive threshold algorithm and then applies it to produce a binary output image. Experimental results show that combined those algorithms is robust in dealing with non-uniform illuminated, low contrast historic document images in terms of both accuracy and efficiency.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127278143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Railway Signal Interlocking Logic Simulation System","authors":"Lidong Zhang","doi":"10.1145/3341016.3341017","DOIUrl":"https://doi.org/10.1145/3341016.3341017","url":null,"abstract":"Railway interlocking systems are apparatuses that prevent conflicting movements of trains through an arrangement of tracks. In this paper, we formulated the main way to design a railway signal interlocking simulation system. To simulate the interlocking logic of railway signal, we first analyzed such devices as signals, track circuit, switches and train routes. Then we designed such classes as signal class, track circuit class, switch class and route class based on object-oriented programming language. By defining the attributes of every class and taking full consideration of the signal relays' types and amounts, we developed the interlocking logic simulation system with C# language. The simulation system is applied on the actual station chart of downward throat and proves it's applicable. The system realized interlocking logic implementation and signal opening functions. Being put into practice, it proves to be worthy of promotion and widely used.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"474 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122739002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}