Linna Li, Zhengying Ma, Dongwang Zhong, Tengfei Li, Qi Zhang
{"title":"Damage Analysis of Urban Comprehensive Pipe Gallery Caused by Internal Gas Explosion Based on HHT","authors":"Linna Li, Zhengying Ma, Dongwang Zhong, Tengfei Li, Qi Zhang","doi":"10.1142/s0218001423540083","DOIUrl":"https://doi.org/10.1142/s0218001423540083","url":null,"abstract":"<p>In this paper, the method of embedding piezoelectric ceramic sensors is used to test the damage of the materials and models of the urban comprehensive pipe gallery. The monitoring signal is processed by HHT method, and the frequency and energy changes of the piezoelectric signal before and after the explosion were analyzed, thus the damage characteristics of the urban comprehensive pipe gallery under the internal gas explosion are determined. The results show that in the gas explosion test inside the pipe gallery model, the waveform and peak value of the piezoelectric signal before and after the explosion did not change significantly, indicating that there was no obvious damage to the side wall of the integrated pipe gallery. However, after further time-frequency analysis of the piezoelectric signal by HHT, it is found that the dominant frequency of the piezoelectric signal after explosion has a downward trend, so it is judged that there is a small damage in the pipe gallery after explosion.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"43 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140629430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sanjay Kumar Sonbhadra, Sonali Agarwal, P. Nagabhushan
{"title":"Pinball-OCSVM for Early-Stage COVID-19 Diagnosis with Limited Posteroanterior Chest X-Ray Images","authors":"Sanjay Kumar Sonbhadra, Sonali Agarwal, P. Nagabhushan","doi":"10.1142/s0218001424570027","DOIUrl":"https://doi.org/10.1142/s0218001424570027","url":null,"abstract":"<p>The conventional way of respiratory coronavirus disease 2019 (COVID-19) diagnosis is reverse transcription polymerase chain reaction (RT-PCR), which is less sensitive during early stages; especially if the patient is asymptomatic, which may further cause more severe pneumonia. In this context, several deep learning models have been proposed to identify pulmonary infections using publicly available chest X-ray (CXR) image datasets for early diagnosis, better treatment and quick cure. In these datasets, the presence of less number of COVID-19 positive samples compared to other classes (normal, pneumonia and Tuberculosis) raises the challenge for unbiased learning of deep learning models. This learning problem can be considered as one-class classification problem where the target class samples are present and other classes are absent or ill-defined. All deep learning models opted class balancing techniques to solve this issue; which however should be avoided in any medical diagnosis process. Moreover, the deep learning models are also data hungry and need massive computation resources. Therefore, for quicker diagnosis, this research proposes a novel pinball loss function based one-class support vector machine (PB-OCSVM), that can work in presence of limited COVID-19 positive CXR samples (target class or class-of-interest (CoI) samples) with objectives to maximize the learning efficiency and to minimize the false predictions. The performance of the proposed model is compared with conventional OCSVM and recent deep learning models, and the experimental results prove that the proposed model outperformed state-of-the-art methods. To validate the robustness of the proposed model, experiments are also performed with noisy CXR images and UCI benchmark datasets.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"19 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140577926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Guohan Jiang, Tianzhen Wang, Dingding Yang, Jingyi You
{"title":"A Domain Variable Prior Based Multi-Style Transfer Network for Data Augmentation of Tidal Stream Turbine Rotor Image Dataset","authors":"Guohan Jiang, Tianzhen Wang, Dingding Yang, Jingyi You","doi":"10.1142/s021800142454003x","DOIUrl":"https://doi.org/10.1142/s021800142454003x","url":null,"abstract":"<p>The style of the underwater images varies according to the region of the sea. However, Tidal Stream Turbine (TST) rotor images captured in the laboratory environment cannot reflect the real underwater environment in image style, resulting in poor generalization of image signal-based fault detection algorithms. Due to the fixed capture position of the camera, the TST rotor image dataset has a high semantic similarity between images, resulting in content loss in conventional image-to-image translation networks. Meanwhile, the one-to-one translation feature in other works cannot meet our requirements. In this work, a Domain Variable Prior-based Multi-style Transfer Network (DVP-MSTN) is proposed to achieve TST rotor image style augmentation. First, the backbone network is trained using a public paired dataset to acquire prior knowledge of domain variable (Knowledge Acquiring, KA). Next, a Multi-domain Transfer Unit (MDT unit) is introduced to enable the conversion of style representations in low-dimensional space. Finally, the prior knowledge is shared to train the MDT unit by fixing the parameters of the backbone network optimized from the KA process (Knowledge Sharing, KS). In addition, an algorithm based on the dark channel of the image is proposed to improve the transfer of low-contrast features. Specifically, a discriminator is used to discriminate the image dark channel to guide the MDT unit to generate low-contrast style representation conditionally. Meanwhile, color loss is employed to preserve the color feature of the image. By controlling the weights of the style code, this method enables control over the image style transfer process, thereby expanding the variety of image styles in the dataset for the purpose of data augmentation.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"213 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140205229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"DAGAN: A GAN Network for Image Denoising of Medical Images Using Deep Learning of Residual Attention Structures","authors":"Guoxiang Tong, Fangning Hu, Hongjun Liu","doi":"10.1142/s0218001424520037","DOIUrl":"https://doi.org/10.1142/s0218001424520037","url":null,"abstract":"<p>Medical images are susceptible to noise and artifacts, so denoising becomes an essential pre-processing technique for further medical image processing stages. We propose a medical image denoising method based on dual-attention mechanism for generative adversarial networks (GANs). The method is based on a GAN model with fused residual structure and introduces a global skip-layer connection structure to balance the learning ability of the shallow and deep networks. The generative network uses a residual module containing channel and spatial attention for efficient extraction of CT image features. The mean square error loss and perceptual loss are introduced to construct a composite loss function to optimize the model loss function, which helps to improve the image generation effect of the model. Experimental results on the LUNA dataset and “the 2016 Low-Dose CT Grand Challenge” dataset show that DAGAN achieves the best results in root mean square error (RMSE), structural similarity (SSIM) and peak signal-to-noise ratio (PSNR) when compared to the state-of-the-art methods. In particular, PSNR reaches 31.2308 dB and 27.5265 dB, SSIM reaches 0.9115 and 0.7895, while RMSE is 0.0082 and 0.0112, respectively. This indicates that our method performs better than the state-of-the-art methods in the task of CT image denoising.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"103 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140127513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Leveraging Sampling Schemes on Skewed Class Distribution to Enhance Male Fertility Detection with Ensemble AI Learners","authors":"Debasmita GhoshRoy, P. A. Alvi, KC Santosh","doi":"10.1142/s0218001424510030","DOIUrl":"https://doi.org/10.1142/s0218001424510030","url":null,"abstract":"<p>Designing effective AI models becomes a challenge when dealing with imbalanced/skewed class distributions in datasets. Addressing this, re-sampling techniques often come into play as potential solutions. In this investigation, we delve into the male fertility dataset, exploring 14 re-sampling approaches to understand their impact on enhancing predictive model performance. The research employs conventional AI learners to gauge male fertility potential. Notably, five ensemble AI learners are studied, their performances are compared, and their results are evaluated using four measurement indices. Through comprehensive comparative analysis, we identify substantial enhancement in model effectiveness. Our findings showcase that the LightGBM model with SMOTE-ENN re-sampling stands out, achieving an efficacy of 96.66% and an F1-Score of 95.60% through 5-fold cross-validation. Interestingly, the CatBoost model, without re-sampling, exhibits strong performance, achieving an efficacy of 86.99% and an F1-Score of 93.02%. Furthermore, we benchmark our approach against state-of-the-art methods in male fertility prediction, particularly highlighting the use of re-sampling techniques like SMOTE and ESLSMOTE. Consequently, our proposed model emerges as a robust and efficient computational framework, promising accurate male fertility prediction.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"32 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-03-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140129600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Identification Method of Unmanned Aerial Vehicle Graphical Control Strategy Based on Cloud Server","authors":"Zhengyu Liu, Zhenbang Cheng, Yu Liu, Qing Jiang","doi":"10.1142/s0218001424500010","DOIUrl":"https://doi.org/10.1142/s0218001424500010","url":null,"abstract":"<p>With the rapid development of unmanned aerial vehicle (UAV) technology, UAV has been widely used in agricultural plant protection, electric power inspection, security patrols, and other fields. However, the control system of the UAV is a complex human–computer interaction system, which requires higher requirements in practical applications. Due to differences in hardware design, software development, and other aspects among different manufacturers, these UAV control systems require high hardware requirements, resulting in a long development cycle. At the same time, in practical applications, due to various reasons, there are equipment failures that are difficult to detect and eliminate in a timely manner. This paper used the UAV graphical control strategy identification method based on cloud server technology, and used the support vector machine (SVM) algorithm to analyze its identification accuracy. The research results showed that when other conditions were the same, the number of researchers and experts who were satisfied with the drone trial effect of the cloud server was 42 and 10, respectively, accounting for 84% and 100%. It indicates that they believe that the cloud server can effectively improve the effectiveness of the drone graphical control strategy recognition method, indicating a positive relationship between the two.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"363 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-02-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140127605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Head Pose Estimation Based on Multi-Level Feature Fusion","authors":"Chunman Yan, Xiao Zhang","doi":"10.1142/s0218001424560020","DOIUrl":"https://doi.org/10.1142/s0218001424560020","url":null,"abstract":"<p>Head Pose Estimation (HPE) has a wide range of applications in computer vision, but still faces challenges: (1) Existing studies commonly use Euler angles or quaternions as pose labels, which may lead to discontinuity problems. (2) HPE does not effectively address regression via rotated matrices. (3) There is a low recognition rate in complex scenes, high computational requirements, etc. This paper presents an improved unconstrained HPE model to address these challenges. First, a rotation matrix form is introduced to solve the problem of unclear rotation labels. Second, a continuous 6D rotation matrix representation is used for efficient and robust direct regression. The RepVGG-A2 lightweight framework is used for feature extraction, and by adding a multi-level feature fusion module and a coordinate attention mechanism with residual connection, to improve the network’s ability to perceive contextual information and pay attention to features. The model’s accuracy was further improved by replacing the network activation function and improving the loss function. Experiments on the BIWI dataset 7:3 dividing the training and test sets show that the average absolute error of HPE for the proposed network model is 2.41. Trained on the dataset 300W_LP and tested on the AFLW2000 and BIWI datasets, the average absolute errors of HPE of the proposed network model are 4.34 and 3.93. The experimental results demonstrate that the improved network has better HPE performance.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"32 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140129687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Optimized Ensemble Machine Learning Approach for Emotion Detection from Thermal Images","authors":"Jayaprakash Katual, Amit Kaul","doi":"10.1142/s0218001424510029","DOIUrl":"https://doi.org/10.1142/s0218001424510029","url":null,"abstract":"<p>Emotions indicate the feelings of the individual which are linked with personal experiences, moods, and affective states. Detection of emotion can be helpful in many fields like maintaining a patient’s psychological well-being, surveillance, driver monitoring, etc. In this paper, an effective machine learning approach has been put forth for emotion detection where an ensemble of three out of five best-performing classifiers has been formed to enhance the classification accuracy. Two deep learning models (AlexNet and ResNet) have been optimally combined with <span><math altimg=\"eq-00001.gif\" display=\"inline\" overflow=\"scroll\"><mi>k</mi></math></span><span></span>-nearest neighbor (KNN). The optimal weights for ensemble weighted averaging of classifiers have been computed with aid of particle swarm optimization (PSO) and genetic algorithm (GA) optimization. The developed framework has been tested on two publicly available datasets. An overall accuracy of above 95% has been achieved on the testing set for both datasets. The best performance was obtained by training the classifiers with segmented images and combining them by using the weights obtained through PSO. The results depicted the efficiency of the optimized ensemble machine learning approach for all performance measures used in this study in comparison to the performance of individual classifiers and majority voting fusion.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"82 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140127696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
G Suresh, G Manikandan, G Bhuvaneswari, P Shanthakumar
{"title":"Pelican Whale Optimization Enabled Deep Learning Framework for Video Steganography Using Arnold Transform-Based Embedding","authors":"G Suresh, G Manikandan, G Bhuvaneswari, P Shanthakumar","doi":"10.1142/s0218001423590267","DOIUrl":"https://doi.org/10.1142/s0218001423590267","url":null,"abstract":"<p>Steganography refers to hiding a secret message from various sources, such as images, videos, audio and so on. The advantage of steganography is to avoid data hacking in transmission medium during the transmission of information sources. Video steganography is superior to image steganography since the videos can hide a substantial quantity of secret messages more than the image. Hence, this research introduced the video stereography technique, Arnold Transform with SqueezeNet-based Pelican Whale Optimization Algorithm (AT<span><math altimg=\"eq-00001.gif\" display=\"inline\" overflow=\"scroll\"><mo stretchy=\"false\">+</mo></math></span><span></span>SqueezeNet_PWOA), for concealing the secret image on the video. To hide the secret image on the video, the proposed method follows three steps: key frame and feature extraction, pixel prediction and embedding. The extraction of the key frame process is carried out by the Structural Similarity Index Measure (SSIM), and then the neighborhood features and convolutional neural network (CNN) features are extracted from the frame to improve the robustness of the embedding process. Moreover, the pixel prediction is completed by the SqueezeNet model, wherein the learning factors are tuned by the PWOA. In addition, the embedding process is completed by applying the Arnold transform on the predicted pixel, and the transformed regions are combined with the secret image using the embedding function. Likewise, the extraction process extracts the secret image from the embedded video by substituting the predicted pixel and Arnold transform on the embedded video. The proposed method is used to hide chunks of secret data in the form of video sequences and it improves the performance. The Arnold transform used in this work provides security by encrypting the data. The use of SqueezeNet makes the proposed model a simple design and this reduces the computational time. Thus, the AT<span><math altimg=\"eq-00002.gif\" display=\"inline\" overflow=\"scroll\"><mo stretchy=\"false\">+</mo></math></span><span></span>SqeezeNet_PWOA attained better correlation coefficient (CC), peak signal-to-noise ratio (PSNR) and mean square error (MSE) of 0.908, 48.66 and 0.001 dB with the Gaussian noise.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"75 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140127597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Neural Network-Based Algorithm for Identification of Recaptured Images","authors":"Changming Liu, Yanjun Sun, Lin Deng, Yan Sun","doi":"10.1142/s0218001423500362","DOIUrl":"https://doi.org/10.1142/s0218001423500362","url":null,"abstract":"<p>With the improvement of digital image display technology, the “secondary imaging” caused by digital cameras is also gradually popularized, and the quality of the recaptured image formed by this imaging is also getting higher and higher, and this kind of high-quality fake image has caused great threat to digital images security. We propose a neural network-based recaptured image identification algorithm and use the difference between two types of images to build the identification algorithm in the frequency domain. The algorithm uses filtering to obtain the feature images which are the high-frequency and low-frequency filtering images, in order to further distinguish the image differences, the direction of the filtered image obtained from high-frequency images, each direction of the filtered image contains high-frequency information at different angles, and the low-frequency image is downsampled. At the same time, the low-frequency image is downsampled to obtain a multi-scale filtered image. The algorithm extracts the features from previous images as the feature values for classification, and finally uses neural networks for classification to obtain the classification results, and these prove that the algorithm presented is able to differentiate the recaptured images effectively in this paper.</p>","PeriodicalId":54949,"journal":{"name":"International Journal of Pattern Recognition and Artificial Intelligence","volume":"143 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140152459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}