AI Communications最新文献

筛选
英文 中文
Sound event localization and detection using element-wise attention gate and asymmetric convolutional recurrent neural networks 基于元素注意门和非对称卷积递归神经网络的声音事件定位与检测
IF 0.8 4区 计算机科学
AI Communications Pub Date : 2023-04-20 DOI: 10.3233/aic-220125
Lean Yan, Min Guo, Zhiqiang Li
{"title":"Sound event localization and detection using element-wise attention gate and asymmetric convolutional recurrent neural networks","authors":"Lean Yan, Min Guo, Zhiqiang Li","doi":"10.3233/aic-220125","DOIUrl":"https://doi.org/10.3233/aic-220125","url":null,"abstract":"There are problems that standard square convolution kernel has insufficient representation ability and recurrent neural network usually ignores the importance of different elements within an input vector in sound event localization and detection. This paper proposes an element-wise attention gate-asymmetric convolutional recurrent neural network (EleAttG-ACRNN), to improve the performance of sound event localization and detection. First, a convolutional neural network with context gating and asymmetric squeeze excitation residual is constructed, where asymmetric convolution enhances the capability of the square convolution kernel; squeeze excitation can improve the interdependence between channels; context gating can weight the important features and suppress the irrelevant features. Next, in order to improve the expressiveness of the model, we integrate the element-wise attention gate into the bidirectional gated recurrent network, which is to highlight the importance of different elements within an input vector, and further learn the temporal context information. Evaluation results using the TAU Spatial Sound Events 2019-Ambisonic dataset show the effectiveness of the proposed method, and it improves SELD performance up to 0.05 in error rate, 1.7% in F-score, 0.7° in DOA error, and 4.5% in Frame recall compared to a CRNN method.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"27 1","pages":"147-157"},"PeriodicalIF":0.8,"publicationDate":"2023-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78478034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The 11th IJCAR automated theorem proving system competition - CASC-J11 第十一届IJCAR自动定理证明系统竞赛——CASC-J11
IF 0.8 4区 计算机科学
AI Communications Pub Date : 2023-03-30 DOI: 10.3233/aic-220244
G. Sutcliffe, Martin Desharnais
{"title":"The 11th IJCAR automated theorem proving system competition - CASC-J11","authors":"G. Sutcliffe, Martin Desharnais","doi":"10.3233/aic-220244","DOIUrl":"https://doi.org/10.3233/aic-220244","url":null,"abstract":"The CADE ATP System Competition (CASC) is the annual evaluation of fully automatic, classical logic, Automated Theorem Proving (ATP) systems. CASC-J11 was the twenty-seventh competition in the CASC series. Twenty-four ATP systems competed in the various competition divisions. This paper presents an outline of the competition design and a commentated summary of the results.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"41 1","pages":"73-91"},"PeriodicalIF":0.8,"publicationDate":"2023-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85831025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automatic Chinese knowledge-based question answering by the MGBA-LSTM-CNN model 基于MGBA-LSTM-CNN模型的中文知识自动问答
IF 0.8 4区 计算机科学
AI Communications Pub Date : 2023-03-13 DOI: 10.3233/aic-210003
Wenyuan Liu, Mingliang Fan, Kai Feng, Dingding Guo
{"title":"Automatic Chinese knowledge-based question answering by the MGBA-LSTM-CNN model","authors":"Wenyuan Liu, Mingliang Fan, Kai Feng, Dingding Guo","doi":"10.3233/aic-210003","DOIUrl":"https://doi.org/10.3233/aic-210003","url":null,"abstract":"The purpose of knowledge-based question answering (KBQA) is to accurately answer the questions raised by users through knowledge triples. Traditional Chinese KBQA methods rely heavily on artificial features, resulting in unsatisfactory QA results. To solve the above problems, this paper divides Chinese KBQA into two parts: entity extraction and attribute mapping. In the entity extraction stage, the improved Bi-LSTM-CNN-CRF model is used to identify the entity of questions and the Levenshtein distance method is used to resolve the entity link error. In the attribute mapping stage, according to the characteristics of questions and candidate attributes, the MGBA-LSTM-CNN model is proposed to encode questions and candidate attributes from the semantic level and word level, respectively, and splice them into new semantic vectors. Finally, the cosine distance is used to measure the similarity of the two vectors to find candidate attributes most similar to questions. The experimental results show that the system achieves good results in the Chinese question and answer data set.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"1 1","pages":"93-110"},"PeriodicalIF":0.8,"publicationDate":"2023-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77832268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PairTraining: A method for training Convolutional Neural Networks with image pairs PairTraining:一种用图像对训练卷积神经网络的方法
IF 0.8 4区 计算机科学
AI Communications Pub Date : 2023-02-10 DOI: 10.3233/aic-220145
Yuhong Shi, Yan Zhao, C. Yao
{"title":"PairTraining: A method for training Convolutional Neural Networks with image pairs","authors":"Yuhong Shi, Yan Zhao, C. Yao","doi":"10.3233/aic-220145","DOIUrl":"https://doi.org/10.3233/aic-220145","url":null,"abstract":"In the field of image classification, the Convolutional Neural Networks (CNNs) are effective. Most of the work focuses on improving and innovating CNN’s network structure. However, using labeled data more effectively for training has also been an essential part of CNN’s research. Combining image disturbance and consistency regularization theory, this paper proposes a model training method (PairTraining) that takes image pairs as input and dynamically modify the training difficulty according to the accuracy of the model in the training set. According to the accuracy of the model in the training set, the training process will be divided into three stages: the qualitative stage, the fine learning stage and the strengthening learning stage. Contrastive learning images are formed using a progressively enhanced image disturbance strategy at different training stages. The input image and contrast learning image are combined into image pairs for model training. The experiments are tested on four public datasets using eleven CNN models. These models have different degrees of improvement in accuracy on the four datasets. PairTraining can adapt to a variety of CNN models for image classification training. This method can better improve the effectiveness of training and improve the degree of generalization of classification models after training. The classification model obtained by PairTraining has better performance in practical application.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"36 1","pages":"111-126"},"PeriodicalIF":0.8,"publicationDate":"2023-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83259888","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning invariant representation using synthetic imagery for object detection 使用合成图像学习不变表示用于目标检测
4区 计算机科学
AI Communications Pub Date : 2023-02-09 DOI: 10.3233/aic-220039
Ning Jiang, Jinglong Fang, Yanli Shao
{"title":"Learning invariant representation using synthetic imagery for object detection","authors":"Ning Jiang, Jinglong Fang, Yanli Shao","doi":"10.3233/aic-220039","DOIUrl":"https://doi.org/10.3233/aic-220039","url":null,"abstract":"Recent years have witnessed a rapid advance in training and testing synthetic data through deep learning networks for the annotation of synthetic data that can be automatically marked. However, a domain discrepancy still exists between synthetic data and real data. In this paper, we address the domain discrepancy issue from three aspects: 1) We design a synthetic image generator with automatically labeled based on 3D scenes. 2) A novel adversarial domain adaptation model is proposed to learn robust intermediate representation free of distractors to improve the transfer performance. 3) We construct a distractor-invariant network and adopt the sample transferability strategy on global-local levels, respectively, to mitigate the cross-domain gap. Additional exploratory experiments demonstrate that the proposed model achieves large performance margins, which show significant advance over the other state-of-the-art models, performing a promotion of 10%–15% mAP on various domain adaptation scenarios.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"332 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136156728","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Defending against adversarial attacks on graph neural networks via similarity property 基于相似性的图神经网络对抗性攻击防御
IF 0.8 4区 计算机科学
AI Communications Pub Date : 2023-01-23 DOI: 10.3233/aic-220120
Minghong Yao, Haizheng Yu, H. Bian
{"title":"Defending against adversarial attacks on graph neural networks via similarity property","authors":"Minghong Yao, Haizheng Yu, H. Bian","doi":"10.3233/aic-220120","DOIUrl":"https://doi.org/10.3233/aic-220120","url":null,"abstract":"Graph Neural Networks (GNNs) are powerful tools in graph application areas. However, recent studies indicate that GNNs are vulnerable to adversarial attacks, which can lead GNNs to easily make wrong predictions for downstream tasks. A number of works aim to solve this problem but what criteria we should follow to clean the perturbed graph is still a challenge. In this paper, we propose GSP-GNN, a general framework to defend against massive poisoning attacks that can perturb graphs. The vital principle of GSP-GNN is to explore the similarity property to mitigate negative effects on graphs. Specifically, this method prunes adversarial edges by the similarity of node feature and graph structure to eliminate adversarial perturbations. In order to stabilize and enhance GNNs training process, previous layer information is adopted in case a large number of edges are pruned in one layer. Extensive experiments on three real-world graphs demonstrate that GSP-GNN achieves significantly better performance compared with the representative baselines and has favorable generalization ability simultaneously.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"25 1","pages":"27-39"},"PeriodicalIF":0.8,"publicationDate":"2023-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87490723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic prediction of epileptic seizure using hybrid deep ResNet-LSTM model 基于混合深度ResNet-LSTM模型的癫痫发作自动预测
IF 0.8 4区 计算机科学
AI Communications Pub Date : 2023-01-23 DOI: 10.3233/aic-220177
Y. Singh, D. K. Lobiyal
{"title":"Automatic prediction of epileptic seizure using hybrid deep ResNet-LSTM model","authors":"Y. Singh, D. K. Lobiyal","doi":"10.3233/aic-220177","DOIUrl":"https://doi.org/10.3233/aic-220177","url":null,"abstract":"Numerous advanced data processing and machine learning techniques for identifying epileptic seizures have been developed in the last two decades. Nonetheless, many of these solutions need massive data sets and intricate computations. Our approach transforms electroencephalogram (EEG) data into the time-frequency domain by utilizing a short-time fourier transform (STFT) and the spectrogram (t-f) images as the input stage of the deep learning model. Using EEG data, we have constructed a hybrid model comprising of a Deep Convolution Network (ResNet50) and a Long Short-Term Memory (LSTM) for predicting epileptic seizures. Spectrogram images are used to train the proposed hybrid model for feature extraction and classification. We analyzed the CHB-MIT scalp EEG dataset. For each preictal period of 5, 15, and 30 minutes, experiments are conducted to evaluate the performance of the proposed model. The experimental results indicate that the proposed model produced the optimum performance with a 5-minute preictal duration. We achieved an average accuracy of 94.5%, the average sensitivity of 93.7%, the f1-score of 0.9376, and the average false positive rate (FPR) of 0.055. Our proposed technique surpassed the random predictor and other current algorithms used for seizure prediction for all patients’ data in the dataset. One can use the effectiveness of our proposed model to help in the early diagnosis of epilepsy and provide early treatment.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"15 1","pages":"57-72"},"PeriodicalIF":0.8,"publicationDate":"2023-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85929328","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Learning invariant representation using synthetic imagery for object detection 使用合成图像学习不变表示用于目标检测
IF 0.8 4区 计算机科学
AI Communications Pub Date : 2022-12-30 DOI: 10.2139/ssrn.4033177
Ning Jiang, Jinglong Fang, Yanli Shao
{"title":"Learning invariant representation using synthetic imagery for object detection","authors":"Ning Jiang, Jinglong Fang, Yanli Shao","doi":"10.2139/ssrn.4033177","DOIUrl":"https://doi.org/10.2139/ssrn.4033177","url":null,"abstract":"Recent years have witnessed a rapid advance in training and testing synthetic data through deep learning networks for the annotation of synthetic data can be automatically marked. However, domain discrepancy still exists between synthetic data and real data. In this paper, we address the domain discrepancy issue from three aspects: 1) We design a synthetic image generator with automatically labeled based on 3d scenes. 2) A novel adversarial domain adaptation model is proposed to learn robust intermediate representation free of distractors to improve the transfer performance. 3) We construct a distractor-invariant network and adopt the sample transferability strategy on global-local levels respectively to mitigate the cross-domain gap. Additional exploratory experiments demonstrate that the proposed model achieves large performance margins, which show significant advance over the other state-of-the-art models, performing a promotion of 10%–15% mAP on various domain adaptation scenarios.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"14 1","pages":"13-25"},"PeriodicalIF":0.8,"publicationDate":"2022-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91108650","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal interaction aware embedding for location-based social networks 基于位置的社交网络的多模态交互感知嵌入
IF 0.8 4区 计算机科学
AI Communications Pub Date : 2022-12-12 DOI: 10.3233/aic-220161
Ruiyun Yu, Kang Yang, Zhihong Wang, Shi Zhen
{"title":"Multimodal interaction aware embedding for location-based social networks","authors":"Ruiyun Yu, Kang Yang, Zhihong Wang, Shi Zhen","doi":"10.3233/aic-220161","DOIUrl":"https://doi.org/10.3233/aic-220161","url":null,"abstract":"Location-based social networks (LBSNs) have greatly promoted the development of the field of human mobility mining. However, the sparsity, multimodality and heterogeneity nature of the user check-in data remains a great concern for learning high-quality user or other entities representations, especially in the downstream application tasks, such as point-of-interest (POI) recommendation. Most existing methods focus on user preference modeling based on sequential POI tags without exploring the interaction between different modalities (e.g., user-user interactions, user-timestamp interactions, user-POI interactions, etc.). To this end, we introduce a multimodal interaction aware embedding framework to generate reliable entity embeddings on the heterogeneous socio-spatial network. At its core, first, multi-modal interaction sub-graph sampling techniques are designed to capture the heterogeneous contexts; then, a self-supervised contrastive learning technique is leveraged to extract intra-modality and inter-modality interactions in a light way. We conduct experiments on the next-POI recommendation tasks based on three real-world datasets. Experimental results demonstrate the superiority of our model over the state-of-the-art embedding learning algorithms.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"1 1","pages":"41-55"},"PeriodicalIF":0.8,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89563516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SECL: Sampling enhanced contrastive learning SECL:抽样增强对比学习
IF 0.8 4区 计算机科学
AI Communications Pub Date : 2022-09-22 DOI: 10.3233/aic-210234
Yixin Tang, Hua Cheng, Yiquan Fang, Tao Cheng
{"title":"SECL: Sampling enhanced contrastive learning","authors":"Yixin Tang, Hua Cheng, Yiquan Fang, Tao Cheng","doi":"10.3233/aic-210234","DOIUrl":"https://doi.org/10.3233/aic-210234","url":null,"abstract":"Instance-level contrastive learning such as SimCLR has been successful as a powerful method for representation learning. However, SimCLR suffers from problems of sampling bias, feature bias and model collapse. A set-level based Sampling Enhanced Contrastive Learning based on SimCLR (SECL) is proposed in this paper. We use the proposed super-sampling method to expand the augmented samples into a contrastive-positive set, which can learn class features of the target sample to reduce the bias. The contrastive-positive set includes Augmentations (the original augmented samples) and Neighbors (the super-sampled samples).We also introduce a samples-correlation strategy to prevent model collapse, where a positive correlation loss or a negative correlation loss is computed to adjust the balance of model’s Alignment and Uniformity. SECL reaches 94.14% classification precision on SST-2 dataset and 89.25% on ARSC dataset. For the multi-class classification task, SECL achieves 90.99% on AGNews dataset. They are all about 1% higher than the precision of SimCLR. Experiments show that the training convergence of SECL is faster, and SECL reduces the risk of bias and model collapse.","PeriodicalId":50835,"journal":{"name":"AI Communications","volume":"34 1","pages":"1-12"},"PeriodicalIF":0.8,"publicationDate":"2022-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75556766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信