Int. J. Intell. Comput. Cybern.最新文献

筛选
英文 中文
Facial expression recognition based on bidirectional gated recurrent units within deep residual network 基于深度残差网络双向门控循环单元的面部表情识别
Int. J. Intell. Comput. Cybern. Pub Date : 2020-11-05 DOI: 10.1108/ijicc-07-2020-0088
Wenjuan Shen, Xiaoling Li
{"title":"Facial expression recognition based on bidirectional gated recurrent units within deep residual network","authors":"Wenjuan Shen, Xiaoling Li","doi":"10.1108/ijicc-07-2020-0088","DOIUrl":"https://doi.org/10.1108/ijicc-07-2020-0088","url":null,"abstract":"Purposerecent years, facial expression recognition has been widely used in human machine interaction, clinical medicine and safe driving. However, there is a limitation that conventional recurrent neural networks can only learn the time-series characteristics of expressions based on one-way propagation information.Design/methodology/approachTo solve such limitation, this paper proposes a novel model based on bidirectional gated recurrent unit networks (Bi-GRUs) with two-way propagations, and the theory of identity mapping residuals is adopted to effectively prevent the problem of gradient disappearance caused by the depth of the introduced network. Since the Inception-V3 network model for spatial feature extraction has too many parameters, it is prone to overfitting during training. This paper proposes a novel facial expression recognition model to add two reduction modules to reduce parameters, so as to obtain an Inception-W network with better generalization.FindingsFinally, the proposed model is pretrained to determine the best settings and selections. Then, the pretrained model is experimented on two facial expression data sets of CK+ and Oulu- CASIA, and the recognition performance and efficiency are compared with the existing methods. The highest recognition rate is 99.6%, which shows that the method has good recognition accuracy in a certain range.Originality/valueBy using the proposed model for the applications of facial expression, the high recognition accuracy and robust recognition results with lower time consumption will help to build more sophisticated applications in real world.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122871397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Hate speech detection in Twitter using hybrid embeddings and improved cuckoo search-based neural networks 基于混合嵌入和改进的基于布谷鸟搜索的神经网络的Twitter仇恨言论检测
Int. J. Intell. Comput. Cybern. Pub Date : 2020-11-03 DOI: 10.1108/ijicc-06-2020-0061
F. E. Ayo, O. Folorunso, F. T. Ibharalu, I. Osinuga
{"title":"Hate speech detection in Twitter using hybrid embeddings and improved cuckoo search-based neural networks","authors":"F. E. Ayo, O. Folorunso, F. T. Ibharalu, I. Osinuga","doi":"10.1108/ijicc-06-2020-0061","DOIUrl":"https://doi.org/10.1108/ijicc-06-2020-0061","url":null,"abstract":"PurposeHate speech is an expression of intense hatred. Twitter has become a popular analytical tool for the prediction and monitoring of abusive behaviors. Hate speech detection with social media data has witnessed special research attention in recent studies, hence, the need to design a generic metadata architecture and efficient feature extraction technique to enhance hate speech detection.Design/methodology/approachThis study proposes a hybrid embeddings enhanced with a topic inference method and an improved cuckoo search neural network for hate speech detection in Twitter data. The proposed method uses a hybrid embeddings technique that includes Term Frequency-Inverse Document Frequency (TF-IDF) for word-level feature extraction and Long Short Term Memory (LSTM) which is a variant of recurrent neural networks architecture for sentence-level feature extraction. The extracted features from the hybrid embeddings then serve as input into the improved cuckoo search neural network for the prediction of a tweet as hate speech, offensive language or neither.FindingsThe proposed method showed better results when tested on the collected Twitter datasets compared to other related methods. In order to validate the performances of the proposed method, t-test and post hoc multiple comparisons were used to compare the significance and means of the proposed method with other related methods for hate speech detection. Furthermore, Paired Sample t-Test was also conducted to validate the performances of the proposed method with other related methods.Research limitations/implicationsFinally, the evaluation results showed that the proposed method outperforms other related methods with mean F1-score of 91.3.Originality/valueThe main novelty of this study is the use of an automatic topic spotting measure based on naïve Bayes model to improve features representation.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123625647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Roman to Gurmukhi Social Media Text Normalization 罗马人到古尔穆克人的社交媒体文本规范化
Int. J. Intell. Comput. Cybern. Pub Date : 2020-10-30 DOI: 10.1108/ijicc-08-2020-0096
J. Kaur, J. Singh
{"title":"Roman to Gurmukhi Social Media Text Normalization","authors":"J. Kaur, J. Singh","doi":"10.1108/ijicc-08-2020-0096","DOIUrl":"https://doi.org/10.1108/ijicc-08-2020-0096","url":null,"abstract":"PurposeNormalization is an important step in all the natural language processing applications that are handling social media text. The text from social media poses a different kind of problems that are not present in regular text. Recently, a considerable amount of work has been done in this direction, but mostly in the English language. People who do not speak English code mixed the text with their native language and posted text on social media using the Roman script. This kind of text further aggravates the problem of normalizing. This paper aims to discuss the concept of normalization with respect to code-mixed social media text, and a model has been proposed to normalize such text.Design/methodology/approachThe system is divided into two phases – candidate generation and most probable sentence selection. Candidate generation task is treated as machine translation task where the Roman text is treated as source language and Gurmukhi text is treated as the target language. Character-based translation system has been proposed to generate candidate tokens. Once candidates are generated, the second phase uses the beam search method for selecting the most probable sentence based on hidden Markov model.FindingsCharacter error rate (CER) and bilingual evaluation understudy (BLEU) score are reported. The proposed system has been compared with Akhar software and RB_R2G system, which are also capable of transliterating Roman text to Gurmukhi. The performance of the system outperforms Akhar software. The CER and BLEU scores are 0.268121 and 0.6807939, respectively, for ill-formed text.Research limitations/implicationsIt was observed that the system produces dialectical variations of a word or the word with minor errors like diacritic missing. Spell checker can improve the output of the system by correcting these minor errors. Extensive experimentation is needed for optimizing language identifier, which will further help in improving the output. The language model also seeks further exploration. Inclusion of wider context, particularly from social media text, is an important area that deserves further investigation.Practical implicationsThe practical implications of this study are: (1) development of parallel dataset containing Roman and Gurmukhi text; (2) development of dataset annotated with language tag; (3) development of the normalizing system, which is first of its kind and proposes translation based solution for normalizing noisy social media text from Roman to Gurmukhi. It can be extended for any pair of scripts. (4) The proposed system can be used for better analysis of social media text. Theoretically, our study helps in better understanding of text normalization in social media context and opens the doors for further research in multilingual social media text normalization.Originality/valueExisting research work focus on normalizing monolingual text. This study contributes towards the development of a normalization system for multilingual text.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124602797","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Motor imagery-based EEG signals classification by combining temporal and spatial deep characteristics 结合时空深度特征的运动图像脑电信号分类
Int. J. Intell. Comput. Cybern. Pub Date : 2020-09-30 DOI: 10.1108/IJICC-07-2020-0077
Xiaoling Li
{"title":"Motor imagery-based EEG signals classification by combining temporal and spatial deep characteristics","authors":"Xiaoling Li","doi":"10.1108/IJICC-07-2020-0077","DOIUrl":"https://doi.org/10.1108/IJICC-07-2020-0077","url":null,"abstract":"In order to improve the weak recognition accuracy and robustness of the classification algorithm for brain-computer interface (BCI), this paper proposed a novel classification algorithm for motor imagery based on temporal and spatial characteristics extracted by using convolutional neural networks (TS-CNN) model.,According to the proposed algorithm, a five-layer neural network model was constructed to classify the electroencephalogram (EEG) signals. Firstly, the author designed a motor imagery-based BCI experiment, and four subjects were recruited to participate in the experiment for the recording of EEG signals. Then, after the EEG signals were preprocessed, the temporal and spatial characteristics of EEG signals were extracted by longitudinal convolutional kernel and transverse convolutional kernels, respectively. Finally, the classification of motor imagery was completed by using two fully connected layers.,To validate the classification performance and efficiency of the proposed algorithm, the comparative experiments with the state-of-the-arts algorithms are applied to validate the proposed algorithm. Experimental results have shown that the proposed TS-CNN model has the best performance and efficiency in the classification of motor imagery, reflecting on the introduced accuracy, precision, recall, ROC curve and F-score indexes.,The proposed TS-CNN model accurately recognized the EEG signals for different tasks of motor imagery, and provided theoretical basis and technical support for the application of BCI control system in the field of rehabilitation exoskeleton.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132515543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Nature-inspired hybrid deep learning for race detection by face shape features 受自然启发的混合深度学习,通过面部形状特征进行种族检测
Int. J. Intell. Comput. Cybern. Pub Date : 2020-07-29 DOI: 10.1108/ijicc-03-2020-0020
Asha Sukumaran, T. Brindha
{"title":"Nature-inspired hybrid deep learning for race detection by face shape features","authors":"Asha Sukumaran, T. Brindha","doi":"10.1108/ijicc-03-2020-0020","DOIUrl":"https://doi.org/10.1108/ijicc-03-2020-0020","url":null,"abstract":"The humans are gifted with the potential of recognizing others by their uniqueness, in addition with more other demographic characteristics such as ethnicity (or race), gender and age, respectively. Over the decades, a vast count of researchers had undergone in the field of psychological, biological and cognitive sciences to explore how the human brain characterizes, perceives and memorizes faces. Moreover, certain computational advancements have been developed to accomplish several insights into this issue.,This paper intends to propose a new race detection model using face shape features. The proposed model includes two key phases, namely. (a) feature extraction (b) detection. The feature extraction is the initial stage, where the face color and shape based features get mined. Specifically, maximally stable extremal regions (MSER) and speeded-up robust transform (SURF) are extracted under shape features and dense color feature are extracted as color feature. Since, the extracted features are huge in dimensions; they are alleviated under principle component analysis (PCA) approach, which is the strongest model for solving “curse of dimensionality”. Then, the dimensional reduced features are subjected to deep belief neural network (DBN), where the race gets detected. Further, to make the proposed framework more effective with respect to prediction, the weight of DBN is fine tuned with a new hybrid algorithm referred as lion mutated and updated dragon algorithm (LMUDA), which is the conceptual hybridization of lion algorithm (LA) and dragonfly algorithm (DA).,The performance of proposed work is compared over other state-of-the-art models in terms of accuracy and error performance. Moreover, LMUDA attains high accuracy at 100th iteration with 90% of training, which is 11.1, 8.8, 5.5 and 3.3% better than the performance when learning percentage (LP) = 50%, 60%, 70%, and 80%, respectively. More particularly, the performance of proposed DBN + LMUDA is 22.2, 12.5 and 33.3% better than the traditional classifiers DCNN, DBN and LDA, respectively.,This paper achieves the objective detecting the human races from the faces. Particularly, MSER feature and SURF features are extracted under shape features and dense color feature are extracted as color feature. As a novelty, to make the race detection more accurate, the weight of DBN is fine tuned with a new hybrid algorithm referred as LMUDA, which is the conceptual hybridization of LA and DA, respectively.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125913761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Deep learning-based approach for segmentation of glioma sub-regions in MRI 基于深度学习的MRI神经胶质瘤亚区分割方法
Int. J. Intell. Comput. Cybern. Pub Date : 2020-07-21 DOI: 10.1108/ijicc-02-2020-0013
Jiten Chaudhary, Rajneesh Rani, A. Kamboj
{"title":"Deep learning-based approach for segmentation of glioma sub-regions in MRI","authors":"Jiten Chaudhary, Rajneesh Rani, A. Kamboj","doi":"10.1108/ijicc-02-2020-0013","DOIUrl":"https://doi.org/10.1108/ijicc-02-2020-0013","url":null,"abstract":"Brain tumor is one of the most dangerous and life-threatening disease. In order to decide the type of tumor, devising a treatment plan and estimating the overall survival time of the patient, accurate segmentation of tumor region from images is extremely important. The process of manual segmentation is very time-consuming and prone to errors; therefore, this paper aims to provide a deep learning based method, that automatically segment the tumor region from MR images.,In this paper, the authors propose a deep neural network for automatic brain tumor (Glioma) segmentation. Intensity normalization and data augmentation have been incorporated as pre-processing steps for the images. The proposed model is trained on multichannel magnetic resonance imaging (MRI) images. The model outputs high-resolution segmentations of brain tumor regions in the input images.,The proposed model is evaluated on benchmark BRATS 2013 dataset. To evaluate the performance, the authors have used Dice score, sensitivity and positive predictive value (PPV). The superior performance of the proposed model is validated by training very popular UNet model in the similar conditions. The results indicate that proposed model has obtained promising results and is effective for segmentation of Glioma regions in MRI at a clinical level.,The model can be used by doctors to identify the exact location of the tumorous region.,The proposed model is an improvement to the UNet model. The model has fewer layers and a smaller number of parameters in comparison to the UNet model. This helps the network to train over databases with fewer images and gives superior results. Moreover, the information of bottleneck feature learned by the network has been fused with skip connection path to enrich the feature map.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123202441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Mobile sensors' patrol path planning in unobservable border region 不可观测边境地区移动传感器巡逻路径规划
Int. J. Intell. Comput. Cybern. Pub Date : 2020-07-07 DOI: 10.1108/ijicc-03-2020-0026
Wichai Pawgasame, K. Wipusitwarakun
{"title":"Mobile sensors' patrol path planning in unobservable border region","authors":"Wichai Pawgasame, K. Wipusitwarakun","doi":"10.1108/ijicc-03-2020-0026","DOIUrl":"https://doi.org/10.1108/ijicc-03-2020-0026","url":null,"abstract":"PurposeThe border control becomes challenging when a protected region is large and there is a limited number of border patrols. This research paper proposes a novel heuristic-based patrol path planning scheme in order to efficiently patrol with resource scarcity.Design/methodology/approachThe trespasser influencing score, which is determined from the environmental characteristics and trespassing statistic of the region, is used as a heuristic for measuring a chance of approaching a trespasser. The patrol plan is occasionally updated with a new trespassing statistic during a border operation. The performance of the proposed patrol path planning scheme was evaluated and compared with other patrol path planning schemes by the empirical experiment under different scenarios.FindingsThe result from the experiment indicates that the proposed patrol planning outperforms other patrol path planning schemes in terms of the trespasser detection rate, when more environment-aware trespassers are in the region.Research limitations/implicationsThe experiment was conducted through simulated agents in simulated environment, which were assumed to mimic real behavior and environment.Originality/valueThis research paper contributes a heuristic-based patrol path planning scheme that applies the environmental characteristics and dynamic statistic of the region, as well as a border surveillance problem model that would be useful for mobile sensor planning in a border surveillance application.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131303820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Exploring compression and parallelization techniques for distribution of deep neural networks over Edge-Fog continuum - a review 探讨边缘雾连续体上深度神经网络分布的压缩和并行化技术
Int. J. Intell. Comput. Cybern. Pub Date : 2020-06-30 DOI: 10.1108/ijicc-04-2020-0038
Azra Nazir, R. N. Mir, Shaima Qureshi
{"title":"Exploring compression and parallelization techniques for distribution of deep neural networks over Edge-Fog continuum - a review","authors":"Azra Nazir, R. N. Mir, Shaima Qureshi","doi":"10.1108/ijicc-04-2020-0038","DOIUrl":"https://doi.org/10.1108/ijicc-04-2020-0038","url":null,"abstract":"PurposeThe trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant. But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015. Thus, it is high time to explore a different approach to tackle this issue, keeping in view the characteristics and needs of the two fields. Processing at the Edge can boost applications with real-time deadlines while complementing security.Design/methodology/approachThis review paper contributes towards three cardinal directions of research in the field of DL for IoT. The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices, forming the realm of the things for IoT. The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques. An appropriate combination of these techniques, including regularization, quantization, and pruning, can aid in building an effective compression pipeline for establishing DL models for IoT use-cases. The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.FindingsDL models are growing deeper with every passing year. Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm. It is realized that a vertically partitioned compressed deep model can handle the trade-off between size, accuracy, communication overhead, bandwidth utilization, and latency but at the expense of an additionally considerable memory footprint. To reduce the memory budget, we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks. However, the critical point between accuracy and size for such models needs further investigation.Originality/valueTo the best of our knowledge, no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum. Besides covering techniques and frameworks that have tried to bring inference to the Edge, the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT. The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115216855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Adaptive task scheduling in IoT using reinforcement learning 使用强化学习的物联网自适应任务调度
Int. J. Intell. Comput. Cybern. Pub Date : 2020-06-30 DOI: 10.1108/ijicc-03-2020-0021
Mohammad Khalid Pandit, R. N. Mir, M. Chishti
{"title":"Adaptive task scheduling in IoT using reinforcement learning","authors":"Mohammad Khalid Pandit, R. N. Mir, M. Chishti","doi":"10.1108/ijicc-03-2020-0021","DOIUrl":"https://doi.org/10.1108/ijicc-03-2020-0021","url":null,"abstract":"The intelligence in the Internet of Things (IoT) can be embedded by analyzing the huge volumes of data generated by it in an ultralow latency environment. The computational latency incurred by the cloud-only solution can be significantly brought down by the fog computing layer, which offers a computing infrastructure to minimize the latency in service delivery and execution. For this purpose, a task scheduling policy based on reinforcement learning (RL) is developed that can achieve the optimal resource utilization as well as minimum time to execute tasks and significantly reduce the communication costs during distributed execution.,To realize this, the authors proposed a two-level neural network (NN)-based task scheduling system, where the first-level NN (feed-forward neural network/convolutional neural network [FFNN/CNN]) determines whether the data stream could be analyzed (executed) in the resource-constrained environment (edge/fog) or be directly forwarded to the cloud. The second-level NN ( RL module) schedules all the tasks sent by level 1 NN to fog layer, among the available fog devices. This real-time task assignment policy is used to minimize the total computational latency (makespan) as well as communication costs.,Experimental results indicated that the RL technique works better than the computationally infeasible greedy approach for task scheduling and the combination of RL and task clustering algorithm reduces the communication costs significantly.,The proposed algorithm fundamentally solves the problem of task scheduling in real-time fog-based IoT with best resource utilization, minimum makespan and minimum communication cost between the tasks.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127872414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Computer-aided diabetic retinopathy diagnostic model using optimal thresholding merged with neural network 最优阈值与神经网络融合的计算机辅助糖尿病视网膜病变诊断模型
Int. J. Intell. Comput. Cybern. Pub Date : 2020-06-29 DOI: 10.1108/ijicc-11-2019-0119
A. Jadhav, Pushpa B. Patil, Sunil Biradar
{"title":"Computer-aided diabetic retinopathy diagnostic model using optimal thresholding merged with neural network","authors":"A. Jadhav, Pushpa B. Patil, Sunil Biradar","doi":"10.1108/ijicc-11-2019-0119","DOIUrl":"https://doi.org/10.1108/ijicc-11-2019-0119","url":null,"abstract":"Diabetic retinopathy (DR) is a central root of blindness all over the world. Though DR is tough to diagnose in starting stages, and the detection procedure might be time-consuming even for qualified experts. Nowadays, intelligent disease detection techniques are extremely acceptable for progress analysis and recognition of various diseases. Therefore, a computer-aided diagnosis scheme based on intelligent learning approaches is intended to propose for diagnosing DR effectively using a benchmark dataset.,The proposed DR diagnostic procedure involves four main steps: (1) image pre-processing, (2) blood vessel segmentation, (3) feature extraction, and (4) classification. Initially, the retinal fundus image is taken for pre-processing with the help of Contrast Limited Adaptive Histogram Equalization (CLAHE) and average filter. In the next step, the blood vessel segmentation is carried out using a segmentation process with optimized gray-level thresholding. Once the blood vessels are extracted, feature extraction is done, using Local Binary Pattern (LBP), Texture Energy Measurement (TEM based on Laws of Texture Energy), and two entropy computations – Shanon's entropy, and Kapur's entropy. These collected features are subjected to a classifier called Neural Network (NN) with an optimized training algorithm. Both the gray-level thresholding and NN is enhanced by the Modified Levy Updated-Dragonfly Algorithm (MLU-DA), which operates to maximize the segmentation accuracy and to reduce the error difference between the predicted and actual outcomes of the NN. Finally, this classification error can correctly prove the efficiency of the proposed DR detection model.,The overall accuracy of the proposed MLU-DA was 16.6% superior to conventional classifiers, and the precision of the developed MLU-DA was 22% better than LM-NN, 16.6% better than PSO-NN, GWO-NN, and DA-NN. Finally, it is concluded that the implemented MLU-DA outperformed state-of-the-art algorithms in detecting DR.,This paper adopts the latest optimization algorithm called MLU-DA-Neural Network with optimal gray-level thresholding for detecting diabetic retinopathy disease. This is the first work utilizes MLU-DA-based Neural Network for computer-aided Diabetic Retinopathy diagnosis.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-06-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127865987","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信