{"title":"Federated Variational Autoencoder for Collaborative Filtering","authors":"Mirko Palato","doi":"10.1109/IJCNN52387.2021.9533358","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533358","url":null,"abstract":"Recommender Systems (RSs) are valuable technologies that help users in their decision-making process. Generally, RSs are designed with the assumption that a central server stores and manages historical users' behaviors. However, users are nowadays more aware of privacy issues leading to a higher demand for privacy-preserving technologies. To cope with this issue, the Federated Learning (FL) paradigm can provide good performance without harming the users' privacy. Some efforts have been devoted to adapt standard collaborative filtering methods (e.g., matrix factorization) into the FL framework in recent years. In this paper, we present a Federated Variational Autoencoder for Collaborative Filtering (FedVAE), which extends the state-of-the-art MultVAE model. Additionally, we propose an adaptive learning rate schedule to accelerate learning. We also discuss the potential privacy-preserving capabilities of FedVAE. An extensive experimental evaluation on five benchmark data sets shows that our proposal can achieve performance close to MultVAE in a reasonable number of iterations. We also empirically demonstrate that the adaptive learning rate guarantees both accelerated learning and good stability.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122061046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Victor Henrique Alves Ribeiro, P. Cavalin, E. Morais
{"title":"A Dynamic Multi-criteria Multi-engine Approach for Text Simplification","authors":"Victor Henrique Alves Ribeiro, P. Cavalin, E. Morais","doi":"10.1109/IJCNN52387.2021.9533365","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533365","url":null,"abstract":"In this work we present a multi-criteria multi-engine approach for text simplification. The main goal is to demonstrate a way to take advantage of a pool of systems, since in the literature several systems have been proposed for the task, and the results have been improving considerably. Note though, that such systems can behave differently, better or worse than the other ones, according to the input. For this reason, in this work we investigate the benefits of exploiting multiple systems at once, in a single-engine, in order to select the most appropriate simplification output from a pool of candidate outputs. In such an engine, a multi-critera decision making approach selects the final output considering simplicity and similarity scores, by comparing the candidates with the input. Results on both the Turk and WikiSmall corpora indicate that the proposed framework is able to balance the trade-off between bilingual evaluation understudy (BLEU), system output against references and against the input sentence (SARI), and Flesch reading ease scores for existing state-of-the art models.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116811789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"CcGL-GAN: Criss-Cross Attention and Global-Local Discriminator Generative Adversarial Networks for text-to-image synthesis","authors":"Xihong Ye, Luanhao Lu","doi":"10.1109/IJCNN52387.2021.9533396","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533396","url":null,"abstract":"Text-to-image synthesis aims to generate a visually realistic image according to a linguistic text description. Visual quality and semantic consistency are two key objectives. Although remarkable progress has been made in improving visual resolutions leveraging Generative Adversarial Networks (GANs), guaranteeing the semantic conformity remains challenging. In this paper, we address it by proposing a novel Criss-Cross Attention and Global-Local Discriminator Generative Adversarial Networks(CcGL-GAN). CcGL-GAN exploits a Criss-Cross Attention mechanism to capture the variation of contextual description, which enables back generators to generate images more efficiently. Moreover, it utilizes Global-Local discriminators to project low-resolution images onto global linguistic representations, and high-resolution images onto local linguistic representations, which ensures that our model narrows the gap between images and descriptions. Experiments conducted on two publicly available datasets, the CUB and Oxford-102, demonstrate the effectiveness of the proposed CcGL-GAN model.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129588000","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-Label Streaming Feature Selection via Class-Imbalance Aware Rough Set","authors":"Yizhang Zou, Xuegang Hu, Peipei Li, Junlong Li","doi":"10.1109/IJCNN52387.2021.9533614","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533614","url":null,"abstract":"Multi-label feature selection aims to select discriminative attributes in multi-label scenario, but most of existing multi-label feature selection methods fail to consider streaming features, i.e. features gradually flow one by one, which is more common in real-world applications. In addition, though there are already some representative works on multi-label streaming feature selection, they fail to tackle the class-imbalance problem, which exists widely in multi-label learning. In fact, class-imbalance will lead to the performance degradation of multi-label learning models. Thus considering class-imbalance problem in multi-label scenario is beneficial to multi-label feature selection because more precise feature evaluation is achieved. Motivated by this, we propose a new rough set named as class-imbalance aware rough set model which can fit class-imbalance problem well. To address streaming features, we construct a novel streaming feature selection framework called SFSCI(Streaming Feature Selection via Class-Imbalance aware rough set), which contains online irrelevancy discarding and online redundancy reduction. Finally, an empirical study on a series of benchmark data sets demonstrates that the proposed method is superior to other state-of-the-art multi-label feature selection methods, including several multi-label streaming feature selection methods.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129864187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Link Prediction with Multiple Structural Attentions in Multiplex Networks","authors":"Shangrong Huang, Quanyu Ma, Chao Yang, Yazhou Yao","doi":"10.1109/IJCNN52387.2021.9533609","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533609","url":null,"abstract":"Many real networks can be viewed as multiplex networks with more than one layers. As different layers are usually not independent from each other, they can provide complementary information in the task of link prediction. In this paper, with the help of attention mechanism, we dig the structural correlations among different layers of the multiplex network as well as the network structural information of the target layer to make more precise link predictions. Specifically, we introduce three different attentions, namely the intra-layer distance/degree attention, the intra-layer neighbourhood attention, and the interlayer structural attention, to calculate both the influence among nodes in the same layer and the link correlations in different layers. Compared with other state-of-the art methods which usually require the information of node attributes or edge types, we only utilize the topological information of the network and thus provide a more general link prediction solution for multiplex network. We conduct comprehensive experiments on several real-world datatsets of different scales. By comparing with the state-of-the-art link prediction algorithms, we show the advantages of our algorithm, and the effectiveness of different attentions. Also, through visual case studies we uncover some intuitions about the relationship between the graph structure and the existence of a link. We make our source code anonymously available at: (will be released after review)","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128779823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Spatial and Temporal Aware Graph Convolutional Network for Flood Forecasting","authors":"Jun Feng, Zhongyi Wang, Yirui Wu, Yuqi Xi","doi":"10.1109/IJCNN52387.2021.9533694","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533694","url":null,"abstract":"Intelligent flood forecasting systems provide an effective means to forecast flood disaster. Accurate flood flow value prediction is a huge challenge since it's influenced by both spatial and temporal relationship among flood factors. Popular deep learning structures like Long Short-Term Memory (LSTM) network lacks abilities of modeling the spatial correlations of hydrological data, thus cannot yield satisfactory prediction results. Moreover, not all the temporal information is always valuable for flood forecasting. In this paper, we proposed a novel spatial and temporal aware Graph Convolution Network (ST-GCN) for flood prediction, which is capable to extract spatial-temporal information from raw flood data. Moreover, a temporal attention mechanism is introduced to weight the importance of different time steps, thus involving global temporal information to improve flood prediction accuracy. Compared with the existing methods, results on two self-collected datasets show that ST-GCN greatly improves the prediction performance.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"130 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128291049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Saurabh Kumar, Siddharth Dangwal, Soumik Adhikary, D. Bhowmik
{"title":"A Quantum Activation Function for Neural Networks: Proposal and Implementation","authors":"Saurabh Kumar, Siddharth Dangwal, Soumik Adhikary, D. Bhowmik","doi":"10.1109/IJCNN52387.2021.9533362","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533362","url":null,"abstract":"A non-linear activation function is an integral component of neural network algorithms used for various tasks such as data classification and pattern recognition. In neu-romorphic/emerging-hardware-based implementations of neural network algorithms, the non-linear activation function is often implemented through dedicated analog electronics. This enables faster execution of the activation function during training and inference of neural networks compared to conventional digital implementation. Here, with a similar motivation, we propose a novel non-linear activation function that can be used in a neural network for data classification. Our activation function can be implemented by taking advantage of the inherent nonlinearity in qubit preparation and SU(2) operation in quantum mechanics. These operations are directly implementable on quantum hardware through single-qubit quantum gates as we show here. In addition, the SU(2) parameters are adjustable here making the activation function adaptable; we adjust the parameters through classical feedback like in a variational algorithm in quantum machine learning. Using our proposed quantum activation function, we report accurate classification using popular machine learning data sets like Fisher's Iris, Wisconsin's Breast Cancer (WBC), Abalone, and MNIST on three different platforms: simulations on a classical computer, simulations on a quantum simulation framework like Qiskit, and experimental implementation on quantum hardware (IBM-Q). Then we use a Bloch-sphere-based approach to intuitively explain how our proposed quantum activation function, with its adaptability, helps in data classification.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128586212","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Relation-Guided Attention Mechanism for Relational Triple Extraction","authors":"Yi Yang, Xueming Li, Xu Li","doi":"10.1109/IJCNN52387.2021.9533950","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533950","url":null,"abstract":"Relational triples are the essential parts of knowledge graphs, which can be usually found in natural language sentences. Relational triple extraction aims to extract all entity pairs with semantic relations from sentences. Recent studies on triple extraction focus on the triple overlap problem where multiple relational triples share single entities or entity pairs in a sentence. Besides, we find sentences may contain implicit relations, and it is challenging for most existing methods to extract implicit relational triples whose relations are implicit in the sentence. In this paper, we propose a relation-guided attention mechanism (RGAM) for relational triple extraction. Firstly, we extract subjects of all possible triples from the sentence, and then identify the corresponding objects under target relations with relation guidance. We utilize relations as prior knowledge instead of regarding relations as classification labels, and apply attention mechanism to obtain fine-grained relation representations, which guide extracted subjects to find the corresponding objects. Our approach (RGAM) can not only learn multiple dependencies in each triple, but also be suitable for extracting implicit relational triples and handling the overlapping triple problem. Extensive experiments show that our model achieves state-of-the-art performance on two public datasets NYT and WebNLG, which demonstrates the effectiveness of our approach.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128700315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Learning Effective Discriminative Features with Differentiable Magnet Loss","authors":"Xiaojing Zhang, Lin Wang, Bo Yang","doi":"10.1109/IJCNN52387.2021.9534158","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9534158","url":null,"abstract":"Neural network optimization relies on the ability of the loss function to learn highly discriminative features. In recent years, Softmax loss has been widely used to train neural network models in various tasks. In order to further enhance the discriminative power of the learned features, Center loss is introduced as an auxiliary function to aid Softmax loss jointly reduce the intra-class variances. In this paper, we propose a novel loss called Differentiable Magnet loss (DML), which can optimize neural nets independently of Softmax loss without joint supervision. This loss offers a more definite convergence target for each class, which not only allows the sample to be close to the homogeneous (intra-class) center but also to stay away from all heterogeneous (inter-class) centers in the feature embedding space. Extensive experimental results demonstrate the superiority of DML in a variety of classification and clustering tasks. Specifically, the 2-D visualization of the learned embedding features by t-SNE effectively proves that our proposed new loss can learn better discriminative representations.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"8 8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124636419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Dilated Convolution-based Denoising Network for Magnetic Resonance Images","authors":"P. C. Tripathi, Soumen Bag","doi":"10.1109/IJCNN52387.2021.9533653","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533653","url":null,"abstract":"Magnetic Resonance Imaging (MRI) are typically corrupted with random noise. This type of noise exhibits the characteristics of Rician distribution in MRI scans. Noise in MRI scans degrades the accuracy of manual and computerized inspection of diseases. Therefore, denoising of MRI images is an indispensable process before the clinical examination of any disease. In this article, we present a novel denoising neural network for MRI images. The proposed network contains a set of dilated convolutions for Rician noise removal. We have used hybrid dilated convolutions to overcome the gridding problem in the network. The residual learning scheme has also been utilized using a set of skip connections. A substantial amount of supervised MRI data has been developed for end-to-end training of the proposed network. Extensive experiments have been performed on synthetic and real MRI datasets to study the effectiveness of the proposed method. The experimental observations indicate that our method not only achieves promising performance but also retains prominent image information effectively.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130303594","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}