Proceedings of the 2nd International Conference on Control and Computer Vision最新文献

筛选
英文 中文
Early Diagnosis of Alzheimer's Disease Using Deep Learning 利用深度学习早期诊断阿尔茨海默病
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341024
Huanhuan Ji, Zhenbing Liu, W. Yan, R. Klette
{"title":"Early Diagnosis of Alzheimer's Disease Using Deep Learning","authors":"Huanhuan Ji, Zhenbing Liu, W. Yan, R. Klette","doi":"10.1145/3341016.3341024","DOIUrl":"https://doi.org/10.1145/3341016.3341024","url":null,"abstract":"Alzheimer's disease (AD) leads to memory loss and impairment, which may cause further symptoms. It affects lives of patients seriously and is not curable, but early confirmation of AD may be helpful to start proper treatment so as to avoid further brain damage. Over the past decades, machine learning methods have been applied to the classification of AD with results based on manually prepared features and a classifier having a multiple-step architecture. Recently, with the development of deep learning, the end to- end process of neural networks has been employed for pattern classification. In this paper, we focus on early diagnosis of AD based on convolutional neural networks (ConvNets) by using magnetic resonance imaging (MRI). Image slices of gray matter and white matter from MRI have been used as the inputs for classification. Ensemble learning methods have been employed after the convolutional operations for improving the classification by combining outputs of deep learning classifiers [27]. Three base ConvNets were designed, implemented, and compared in this paper. Our method was evaluated based on a dataset from the Alzheimer's Disease Neuroimaging Initiative for the early diagnosis of this illness. In particular, the accuracy rates of our classifications have reached up to 97.65% for AD/mild cognitive impairment and 88.37% for mild cognitive impairment/normal control.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129011951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Deep Residual Network for Single Image Super-Resolution 单幅图像超分辨率的深度残差网络
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341030
Haimin Wang, K. Liao, B. Yan, Run Ye
{"title":"Deep Residual Network for Single Image Super-Resolution","authors":"Haimin Wang, K. Liao, B. Yan, Run Ye","doi":"10.1145/3341016.3341030","DOIUrl":"https://doi.org/10.1145/3341016.3341030","url":null,"abstract":"This paper proposes a Deep Residual Network for Single Image Super-Resolution (DRSR). We build a deep model using residual units that remove unnecessary modules. We can build deeper network at the same computing resources with the modified residual units. Experiments shows that deepening the network structure can fully utilize the image contextual information to improve the image reconstruction quality. The network learns both global residuals and local residuals, making the network easier to train. Our network directly extracts features from Low-Resolution (LR) images to reconstruct High-Resolution (HR) images. Computational complexity of the network is dramatically reduced in this way. Experiments shows that our network not only performs well in subjective visual effect but also achieves a high level in objective evaluation index.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121319189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Using Machine Vision to Command a 6-axis Robot Arm to Act on a Randomly Placed Zinc Die Cast Product 用机器视觉控制六轴机械臂对随机放置的锌压铸产品的动作
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341018
Luke Butters, Zezhong Xu, R. Klette
{"title":"Using Machine Vision to Command a 6-axis Robot Arm to Act on a Randomly Placed Zinc Die Cast Product","authors":"Luke Butters, Zezhong Xu, R. Klette","doi":"10.1145/3341016.3341018","DOIUrl":"https://doi.org/10.1145/3341016.3341018","url":null,"abstract":"The paper presents a method, documenting the machine vision techniques required, to automate a manual process at a local New Zealand manufacturing company. The system was required to monitor a conveyor belt for zinc die cast outputs. A robot arm was informed how to pick up the object. Die cast outputs appear face up or down and can be rotated 360°. Four experiments were conducted to determine how accurately the proposed system could detect if the cast was lying face up or down, determine the robot picking location, and determine the angle of direction while performing an error checking function. Circle Hough transform and principal component analysis, along with other object analysis techniques, were used on 2D datasets. Point cloud data was analysed to determine if the cast was face up or down when taken from a time-of-flight camera.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"386 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126735663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Design Therapy for Post-Stroke Patients with Robotics Tools and Principles of Mirror Neurons Using qEEG Parameter Analysis 基于qEEG参数分析的镜像神经元原理和机器人工具设计脑卒中后患者的治疗
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341027
R. Setiawan, Odilia Valentine, H. Zakaria
{"title":"Design Therapy for Post-Stroke Patients with Robotics Tools and Principles of Mirror Neurons Using qEEG Parameter Analysis","authors":"R. Setiawan, Odilia Valentine, H. Zakaria","doi":"10.1145/3341016.3341027","DOIUrl":"https://doi.org/10.1145/3341016.3341027","url":null,"abstract":"Stroke causes neurological disorders such as reduced muscle motor skills, as well as cognitive, visual, and coordination functions, significantly. The reduced level of independence and mobility of a person can affect their quality of life. So, with rehabilitation program, it is expected that motor skills and cognitive function of stroke patients can be restored. This study focused on designing an integrated therapeutic device by using stationary cycles with the principle of mirror neurons and equipped with qEEG signal analysis for parameters comparison before and after therapy. Channels EEG used are F3, F4, C3, Cz, C4, P3, Pz, and P4 according to the rules of localization 10-20 Ref. The device also designed as a robotic system controlled by the movement intentions detected from EEG signals before the actual movement occurs. The Event-Related Potential (ERP) phenomenon mu waves were used to recognize the movement intentions. The device has been tested on normal subjects. To train the movement intention algorithm, the subject were asked to cycle and stop at prescribed time by audio command. qEEG parameters were displayed through a monitor for realtime observation by physical therapist. This research is part of the design process of the stroke therapy system which will then be tested in several stroke patients. Therapy programs for post-stroke patients using this device are expected to increase mobility of the affected limb by stroke.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115435715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Video Object Segmentation with 3D Convolution Network 基于三维卷积网络的视频对象分割
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341031
Huiyun Tang, Pin Tao, Rui Ma, Yuanchun Shi
{"title":"Video Object Segmentation with 3D Convolution Network","authors":"Huiyun Tang, Pin Tao, Rui Ma, Yuanchun Shi","doi":"10.1145/3341016.3341031","DOIUrl":"https://doi.org/10.1145/3341016.3341031","url":null,"abstract":"We explore a novel method to realize semi-supervised video object segmentation with special spatiotemporal feature extracting structure. Considering 3-dimension convolution network can convolute a volume of image sequence, it is a distinct way to get both spatial and temporal information. Our network is composed of three parts, the visual module, the motion module and the decoder module. The visual module learns object appearance feature from object in the first frame for network to detect specific object in following image sequences. The motion module aims to get spatiotemporal information of image sequences with 3-dimension convolution network, which learns diversities of object temporal appearance and location. The purpose of decoder module is to get foreground object mask from output of visual module and motion module with concatenation and upsampling structure. We evaluate our model on DAVIS segmentation dataset[15]. Our model doesn't need online training compared with most detection-based methods because of visual module. As a result, it takes 0.14 second per frame to get mask which is 71 times faster than the state-of-art method-OSVOS[2]. Our model also shows better performance than most methods proposed in recent years and its meanIOU accuracy is comparable with state-of-art methods.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128561012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Recognizing Faces in Shades of Gray 识别灰色阴影中的面孔
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341029
Alaa E. Abdel-Hakim, M. El-Melegy, Shreen K. Refaay
{"title":"Recognizing Faces in Shades of Gray","authors":"Alaa E. Abdel-Hakim, M. El-Melegy, Shreen K. Refaay","doi":"10.1145/3341016.3341029","DOIUrl":"https://doi.org/10.1145/3341016.3341029","url":null,"abstract":"Face recognition depends on relatively few distinguishing features, when compared with common facial features. This gives color information greater value to recognition and identification processes. However, dealing with grayscale facial images is a must in some cases, e.g. legacy images. In this paper, we investigate the effect of losing color information on face recognition. We propose a novel framework, which utilizes CNN-based colorization before a CNN classifier. The proposed framework is tested on LFW benchmark dataset. The evaluation results prove the success of the proposed framework in reducing the negative effect of dropping the color information on face recognition performance.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117015460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-focus Color Image Fusion using Laplacian Filter and Discrete Fourier Transformation with Qualitative Error Image Metrics 基于拉普拉斯滤波和离散傅立叶变换的多焦点彩色图像融合与定性误差图像度量
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341019
S. Khan, Muzammil Khan, Qiong Ran
{"title":"Multi-focus Color Image Fusion using Laplacian Filter and Discrete Fourier Transformation with Qualitative Error Image Metrics","authors":"S. Khan, Muzammil Khan, Qiong Ran","doi":"10.1145/3341016.3341019","DOIUrl":"https://doi.org/10.1145/3341016.3341019","url":null,"abstract":"The Multi-focus image is limited Depth-of-Field (DOF) of an imaging system causes blur images when the sample is wider than the DOF of the optical system. Color multi-focus image fusion allows merging two multi-focus images and produces a composite image integrating complementary information to better understand the entire scene. This paper introduced a new two-level approach for color image fusion. In the first level, the multi-focus image is enhanced by Laplacian Filter (LF) technique with Discrete Fourier Transform (DFT). In the second level, the enhanced images are further processed by Stationary Wavelet Transform (SWT) and construct a more informative fused image. The experimental results on the color multi-focused images showed the capabilities and improved results of the proposed approach as compared with the traditional SWT method. The output image is evaluated using both a qualitative and the quantitative approach. The standard deviation, mean and entropy quantitative metrics are used to assess the performance of the proposed method. The article also introduced a new type of qualitative performance metric named as \"Qualitative Error Image (QEI)\" to evaluate the proposed method and assess future evaluation.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"333 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132335450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Proposal of Environment Adaptive Software 环境自适应软件的构想
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341035
Y. Yamato, Hirofumi Noguchi, M. Kataoka, Takuma Isoda
{"title":"Proposal of Environment Adaptive Software","authors":"Y. Yamato, Hirofumi Noguchi, M. Kataoka, Takuma Isoda","doi":"10.1145/3341016.3341035","DOIUrl":"https://doi.org/10.1145/3341016.3341035","url":null,"abstract":"Recently, heterogeneous hardware such as GPU and FPGA is used in many systems and also IoT devices are increased repidly. However, to utilize heterogeneous hardware, the hurdles are currently high because of much technical skills. In order to break down such a situation, we think it is required in the future that application programmers only need to write logics to be processed, then software will adapt to the environments with heterogeneous hardware, to make it easy to utilize heterogeneous hardware and IoT devices. Therefore, in this paper, we propose environment adaptive software to operate an once written application with high performance by automatically converting the code and configuring setting so that we can utilize GPU, FPGA and IoT devices in the location to be deployed. We explain a processing flow and elemental technologies to achieve environment adaptive software. We also describe the details of elemental technologies such as automatic GPU offloading which are already under considered.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128064716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Chinese Rubbing Image Binarization based on Deep Learning for Image Denoising 基于深度学习的中文摩擦图像二值化去噪
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341023
Zhi-Kai Huang, Zhen-Ning Wang, Jun-Mei Xi, Ling-Ying Hou
{"title":"Chinese Rubbing Image Binarization based on Deep Learning for Image Denoising","authors":"Zhi-Kai Huang, Zhen-Ning Wang, Jun-Mei Xi, Ling-Ying Hou","doi":"10.1145/3341016.3341023","DOIUrl":"https://doi.org/10.1145/3341016.3341023","url":null,"abstract":"Aiming at the problem of the chinese rubbing image segmentation under a denoising algorithm based on deep convolutional neural network is proposed. Document enhancement and binarization is the main pre-processing step in document analysis process. At first, a feed-forward denoising convolutional neural networks as a pre-processing methods for document image has been used for denoise images of additive white Gaussian noise(AWGN). The residual learning mechanism is used to learn the mapping from the noisy image to the residual image between the noisy image and the clean image in the neural network training process. A median filtering has been employed for denoising 'salt and pepper' noise. Given the learned denoising and enhanced image, we compute the adaptive threshold image using local adaptive threshold algorithm and then applies it to produce a binary output image. Experimental results show that combined those algorithms is robust in dealing with non-uniform illuminated, low contrast historic document images in terms of both accuracy and efficiency.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127278143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Railway Signal Interlocking Logic Simulation System 铁路信号联锁逻辑仿真系统
Proceedings of the 2nd International Conference on Control and Computer Vision Pub Date : 2019-06-15 DOI: 10.1145/3341016.3341017
Lidong Zhang
{"title":"Railway Signal Interlocking Logic Simulation System","authors":"Lidong Zhang","doi":"10.1145/3341016.3341017","DOIUrl":"https://doi.org/10.1145/3341016.3341017","url":null,"abstract":"Railway interlocking systems are apparatuses that prevent conflicting movements of trains through an arrangement of tracks. In this paper, we formulated the main way to design a railway signal interlocking simulation system. To simulate the interlocking logic of railway signal, we first analyzed such devices as signals, track circuit, switches and train routes. Then we designed such classes as signal class, track circuit class, switch class and route class based on object-oriented programming language. By defining the attributes of every class and taking full consideration of the signal relays' types and amounts, we developed the interlocking logic simulation system with C# language. The simulation system is applied on the actual station chart of downward throat and proves it's applicable. The system realized interlocking logic implementation and signal opening functions. Being put into practice, it proves to be worthy of promotion and widely used.","PeriodicalId":278141,"journal":{"name":"Proceedings of the 2nd International Conference on Control and Computer Vision","volume":"474 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122739002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信