{"title":"Automatic detection of indoor occupancy based on improved YOLOv5 model.","authors":"Chao Wang, Yunchu Zhang, Yanfei Zhou, Shaohan Sun, Hanyuan Zhang, Yepeng Wang","doi":"10.1007/s00521-022-07730-3","DOIUrl":"https://doi.org/10.1007/s00521-022-07730-3","url":null,"abstract":"<p><p>Indoor occupancy detection is essential for energy efficiency control and Coronavirus Disease 2019 traceability. The number and location of people can be accurately identified and determined through classroom surveillance video analysis. This information is used to manage environmental equipment such as HVAC and lighting systems to reduce energy use. However, the mainstream one-stage YOLO algorithm still uses an anchor-based mechanism and couples detection heads to predict. This results in slow model convergence and poor detection performance for densely occluded targets. Therefore, this paper proposed a novel decoupled anchor-free VariFocal loss convolutional network algorithm DFV-YOLOv5 for occupancy detection to tackle these problems. The proposed method uses the YOLOv5 algorithm as a baseline. It uses the anchor-free mechanism to reduce the number of design parameters needing heuristic tuning. Afterwards, to reduce the coupling of the model, speed up the model's convergence ability, and improve the model detection performance, the detection head is decoupled based on the YOLOv5 model. It can resolve the conflict between classification and regression tasks. In addition, we use the VariFocal loss to assign more weights to difficult data points to optimize the class imbalance problem and use the training target <i>q</i> to measure positive samples, treating positive and negative samples asymmetrically. The total loss function is redesigned, the <math><msub><mi>L</mi> <mn>1</mn></msub> </math> loss is increased, and the ablation experiment verifies the effect of the improved loss. By applying a hybrid activation function of the sigmoid linear unit and rectified linear unit, we improved the model's nonlinear representation and reduced the model's inference time. Finally, a classroom dataset was constructed to validate the occupancy detection performance of the model. The proposed model was compared with mainstream target detection models regarding average mean precision, memory allocation, execution time, and the number of parameters on the VOC2012, CrowdHuman and self-built datasets. The experimental results show that the method significantly improves the detection accuracy and robustness, shortens the inference time, and proves the practicality of the algorithm in occupancy detection compared with the mainstream target detection model and related variants of the model.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 3","pages":"2575-2599"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9436742/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10592134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A hybrid DNN-LSTM model for detecting phishing URLs.","authors":"Alper Ozcan, Cagatay Catal, Emrah Donmez, Behcet Senturk","doi":"10.1007/s00521-021-06401-z","DOIUrl":"10.1007/s00521-021-06401-z","url":null,"abstract":"<p><p>Phishing is an attack targeting to imitate the official websites of corporations such as banks, e-commerce, financial institutions, and governmental institutions. Phishing websites aim to access and retrieve users' important information such as personal identification, social security number, password, e-mail, credit card, and other account information. Several anti-phishing techniques have been developed to cope with the increasing number of phishing attacks so far. Machine learning and particularly, deep learning algorithms are nowadays the most crucial techniques used to detect and prevent phishing attacks because of their strong learning abilities on massive datasets and their state-of-the-art results in many classification problems. Previously, two types of feature extraction techniques [i.e., character embedding-based and manual natural language processing (NLP) feature extraction] were used in isolation. However, researchers did not consolidate these features and therefore, the performance was not remarkable. Unlike previous works, our study presented an approach that utilizes both feature extraction techniques. We discussed how to combine these feature extraction techniques to fully utilize from the available data. This paper proposes hybrid deep learning models based on long short-term memory and deep neural network algorithms for detecting phishing uniform resource locator and evaluates the performance of the models on phishing datasets. The proposed hybrid deep learning models utilize both character embedding and NLP features, thereby simultaneously exploiting deep connections between characters and revealing NLP-based high-level connections. Experimental results showed that the proposed models achieve superior performance than the other phishing detection models in terms of accuracy metric.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 7","pages":"4957-4973"},"PeriodicalIF":4.5,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8349600/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10703149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Human activity recognition from sensor data using spatial attention-aided CNN with genetic algorithm.","authors":"Apu Sarkar, S K Sabbir Hossain, Ram Sarkar","doi":"10.1007/s00521-022-07911-0","DOIUrl":"https://doi.org/10.1007/s00521-022-07911-0","url":null,"abstract":"<p><p>Capturing time and frequency relationships of time series signals offers an inherent barrier for automatic human activity recognition (HAR) from wearable sensor data. Extracting spatiotemporal context from the feature space of the sensor reading sequence is challenging for the current recurrent, convolutional, or hybrid activity recognition models. The overall classification accuracy also gets affected by large size feature maps that these models generate. To this end, in this work, we have put forth a hybrid architecture for wearable sensor data-based HAR. We initially use Continuous Wavelet Transform to encode the time series of sensor data as multi-channel images. Then, we utilize a Spatial Attention-aided Convolutional Neural Network (CNN) to extract higher-dimensional features. To find the most essential features for recognizing human activities, we develop a novel feature selection (FS) method. In order to identify the fitness of the features for the FS, we first employ three filter-based methods: Mutual Information (MI), Relief-F, and minimum redundancy maximum relevance (mRMR). The best set of features is then chosen by removing the lower-ranked features using a modified version of the Genetic Algorithm (GA). The K-Nearest Neighbors (KNN) classifier is then used to categorize human activities. We conduct comprehensive experiments on five well-known, publicly accessible HAR datasets, namely UCI-HAR, WISDM, MHEALTH, PAMAP2, and HHAR. Our model significantly outperforms the state-of-the-art models in terms of classification performance. We also observe an improvement in overall recognition accuracy with the use of GA-based FS technique with a lower number of features. The source code of the paper is publicly available here https://github.com/apusarkar2195/HAR_WaveletTransform_SpatialAttention_FeatureSelection.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 7","pages":"5165-5191"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9596348/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10757508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Prabal Datta Barua, Emrah Aydemir, Sengul Dogan, Mehmet Erten, Feyzi Kaysi, Turker Tuncer, Hamido Fujita, Elizabeth Palmer, U Rajendra Acharya
{"title":"Novel favipiravir pattern-based learning model for automated detection of specific language impairment disorder using vowels.","authors":"Prabal Datta Barua, Emrah Aydemir, Sengul Dogan, Mehmet Erten, Feyzi Kaysi, Turker Tuncer, Hamido Fujita, Elizabeth Palmer, U Rajendra Acharya","doi":"10.1007/s00521-022-07999-4","DOIUrl":"10.1007/s00521-022-07999-4","url":null,"abstract":"<p><p>Specific language impairment (SLI) is one of the most common diseases in children, and early diagnosis can help to obtain better timely therapy economically. It is difficult and time-consuming for clinicians to accurately detect SLI through standard clinical assessments. Hence, machine learning algorithms have been developed to assist in the accurate diagnosis of SLI. This work aims to investigate the graph of the favipiravir molecule-based feature extraction function and propose an accurate SLI detection model using vowels. We proposed a novel handcrafted machine learning framework. This architecture comprises the favipiravir molecular structure pattern, statistical feature extractor, wavelet packet decomposition (WPD), iterative neighborhood component analysis (INCA), and support vector machine (SVM) classifier. Two feature extraction models, statistical and textural, are employed in the handcrafted feature generation methodology. A new nature-inspired graph-based feature extractor that uses the chemical depiction of the favipiravir (favipiravir became popular with the COVID-19 pandemic) is employed for feature extraction. Finally, the proposed favipiravir pattern, statistical feature extractor, and wavelet packet decomposition are used to create a feature vector. Moreover, a statistical feature extractor is used in this work. The WPD generates multilevel features, and the most meaningful features are selected using the NCA feature selector. Finally, these chosen features are fed to SVM classifier for automated classification. Two validation methods, (i) leave one subject out (LOSO) and (ii) tenfold cross-validations (CV), are used to obtain robust classification results. Our proposed favipiravir pattern-based model developed using a vowel dataset can detect SLI children with an accuracy of 99.87% and 98.86% using tenfold and LOSO CV strategies, respectively. These results demonstrated the high vowel classification ability of the proposed favipiravir pattern-based model.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 8","pages":"6065-6077"},"PeriodicalIF":4.5,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9660223/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10801917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Performance analysis and comparison of Machine Learning and LoRa-based Healthcare model.","authors":"Navneet Verma, Sukhdip Singh, Devendra Prasad","doi":"10.1007/s00521-023-08411-5","DOIUrl":"10.1007/s00521-023-08411-5","url":null,"abstract":"<p><p>Diabetes Mellitus (DM) is a widespread condition that is one of the main causes of health disasters around the world, and health monitoring is one of the sustainable development topics. Currently, the Internet of Things (IoT) and Machine Learning (ML) technologies work together to provide a reliable method of monitoring and predicting Diabetes Mellitus. In this paper, we present the performance of a model for patient real-time data collection that employs the Hybrid Enhanced Adaptive Data Rate (HEADR) algorithm for the Long-Range (LoRa) protocol of the IoT. On the Contiki Cooja simulator, the LoRa protocol's performance is measured in terms of high dissemination and dynamic data transmission range allocation. Furthermore, by employing classification methods for the detection of diabetes severity levels on acquired data via the LoRa (HEADR) protocol, Machine Learning prediction takes place. For prediction, a variety of Machine Learning classifiers are employed, and the final results are compared with the already existing models where the Random Forest and Decision Tree classifiers outperform the others in terms of precision, recall, <i>F</i>-measure, and receiver operating curve (ROC) in the Python programming language. We also discovered that using <i>k</i>-fold cross-validation on <i>k</i>-neighbors, Logistic regression (LR), and Gaussian Nave Bayes (GNB) classifiers boosted the accuracy.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 17","pages":"12751-12761"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9989556/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9479074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Malak Abdullah, Mahmoud Al-Ayyoub, Saif AlRawashdeh, Farah Shatnawi
{"title":"E-learningDJUST: E-learning dataset from Jordan university of science and technology toward investigating the impact of COVID-19 pandemic on education.","authors":"Malak Abdullah, Mahmoud Al-Ayyoub, Saif AlRawashdeh, Farah Shatnawi","doi":"10.1007/s00521-021-06712-1","DOIUrl":"10.1007/s00521-021-06712-1","url":null,"abstract":"<p><p>Recently, the COVID-19 pandemic has triggered different behaviors in education, especially during the lockdown, to contain the virus outbreak in the world. As a result, educational institutions worldwide are currently using online learning platforms to maintain their education presence. This research paper introduces and examines a dataset, E-LearningDJUST, that represents a sample of the student's study progress during the pandemic at Jordan University of Science and Technology (JUST). The dataset depicts a sample of the university's students as it includes 9,246 students from 11 faculties taking four courses in spring 2020, summer 2020, and fall 2021 semesters. To the best of our knowledge, it is the first collected dataset that reflects the students' study progress within a Jordanian institute using e-learning system records. One of this work's key findings is observing a high correlation between e-learning events and the final grades out of 100. Thus, the E-LearningDJUST dataset has been experimented with two robust machine learning models (Random Forest and XGBoost) and one simple deep learning model (Feed Forward Neural Network) to predict students' performances. Using RMSE as the primary evaluation criteria, the RMSE values range between 7 and 17. Among the other main findings, the application of feature selection with the random forest leads to better prediction results for all courses as the RMSE difference ranges between (0-0.20). Finally, a comparison study examined students' grades before and after the Coronavirus pandemic to understand how it impacted their grades. A high success rate has been observed during the pandemic compared to what it was before, and this is expected because the exams were online. However, the proportion of students with high marks remained similar to that of pre-pandemic courses.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 16","pages":"11481-11495"},"PeriodicalIF":4.5,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8590139/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9492167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Aidan Cousins, Lucas Nakano, Emma Schofield, Rasa Kabaila
{"title":"A neural network approach to optimising treatments for depression using data from specialist and community psychiatric services in Australia, New Zealand and Japan.","authors":"Aidan Cousins, Lucas Nakano, Emma Schofield, Rasa Kabaila","doi":"10.1007/s00521-021-06710-3","DOIUrl":"https://doi.org/10.1007/s00521-021-06710-3","url":null,"abstract":"<p><p>This study investigated the application of a recurrent neural network for optimising pharmacological treatment for depression. A clinical dataset of 458 participants from specialist and community psychiatric services in Australia, New Zealand and Japan were extracted from an existing custom-built, web-based tool called <i>Psynary</i> . This data, which included baseline and self-completed reviews, was used to train and refine a novel algorithm which was a fully connected network feature extractor and long short-term memory algorithm was firstly trained in isolation and then integrated and annealed using slow learning rates due to the low dimensionality of the data. The accuracy of predicting depression remission before processing patient review data was 49.8%. After processing only 2 reviews, the accuracy was 76.5%. When considering a change in medication, the precision of changing medications was 97.4% and the recall was 71.4% . The medications with predicted best results were antipsychotics (88%) and selective serotonin reuptake inhibitors (87.9%). <i>This is the first study that has created an all-in-one algorithm for optimising treatments for all subtypes of depression.</i> Reducing treatment optimisation time for patients suffering with depression may lead to earlier remission and hence reduce the high levels of disability associated with the condition. Furthermore, in a setting where mental health conditions are increasing strain on mental health services, the utilisation of web-based tools for remote monitoring and machine/deep learning algorithms may assist clinicians in both specialist and primary care in extending specialist mental healthcare to a larger patient community.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 16","pages":"11497-11516"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8754538/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9503950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Toward real-time and efficient cardiovascular monitoring for COVID-19 patients by 5G-enabled wearable medical devices: a deep learning approach.","authors":"Liang Tan, Keping Yu, Ali Kashif Bashir, Xiaofan Cheng, Fangpeng Ming, Liang Zhao, Xiaokang Zhou","doi":"10.1007/s00521-021-06219-9","DOIUrl":"10.1007/s00521-021-06219-9","url":null,"abstract":"<p><p>Patients with deaths from COVID-19 often have co-morbid cardiovascular disease. Real-time cardiovascular disease monitoring based on wearable medical devices may effectively reduce COVID-19 mortality rates. However, due to technical limitations, there are three main issues. First, the traditional wireless communication technology for wearable medical devices is difficult to satisfy the real-time requirements fully. Second, current monitoring platforms lack efficient streaming data processing mechanisms to cope with the large amount of cardiovascular data generated in real time. Third, the diagnosis of the monitoring platform is usually manual, which is challenging to ensure that enough doctors online to provide a timely, efficient, and accurate diagnosis. To address these issues, this paper proposes a 5G-enabled real-time cardiovascular monitoring system for COVID-19 patients using deep learning. Firstly, we employ 5G to send and receive data from wearable medical devices. Secondly, Flink streaming data processing framework is applied to access electrocardiogram data. Finally, we use convolutional neural networks and long short-term memory networks model to obtain automatically predict the COVID-19 patient's cardiovascular health. Theoretical analysis and experimental results show that our proposal can well solve the above issues and improve the prediction accuracy of cardiovascular disease to 99.29%.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 19","pages":"13921-13934"},"PeriodicalIF":6.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/s00521-021-06219-9","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9526794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Machine learning-based diffusion model for prediction of coronavirus-19 outbreak.","authors":"Supriya Raheja, Shreya Kasturia, Xiaochun Cheng, Manoj Kumar","doi":"10.1007/s00521-021-06376-x","DOIUrl":"10.1007/s00521-021-06376-x","url":null,"abstract":"<p><p>The coronavirus pandemic has been globally impacting the health and prosperity of people. A persistent increase in the number of positive cases has boost the stress among governments across the globe. There is a need of approach which gives more accurate predictions of outbreak. This paper presents a novel approach called diffusion prediction model for prediction of number of coronavirus cases in four countries: India, France, China and Nepal. Diffusion prediction model works on the diffusion process of the human contact. Model considers two forms of spread: when the spread takes time after infecting one person and when the spread is immediate after infecting one person. It makes the proposed model different over other state-of-the art models. It is giving more accurate results than other state-of-the art models. The proposed diffusion prediction model forecasts the number of new cases expected to occur in next 4 weeks. The model has predicted the number of confirmed cases, recovered cases, deaths and active cases. The model can facilitate government to be well prepared for any abrupt rise in this pandemic. The performance is evaluated in terms of accuracy and error rate and compared with the prediction results of support vector machine, logistic regression model and convolution neural network. The results prove the efficiency of the proposed model.</p>","PeriodicalId":49766,"journal":{"name":"Neural Computing & Applications","volume":"35 19","pages":"13755-13774"},"PeriodicalIF":4.5,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8358916/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9526796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}