2021 International Joint Conference on Neural Networks (IJCNN)最新文献

筛选
英文 中文
Layerwise Approximate Inference for Bayesian Uncertainty Estimates on Deep Neural Networks 深度神经网络贝叶斯不确定性估计的分层近似推理
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9534229
Ni Zhang, Xiaoyi Chen, Li Quan
{"title":"Layerwise Approximate Inference for Bayesian Uncertainty Estimates on Deep Neural Networks","authors":"Ni Zhang, Xiaoyi Chen, Li Quan","doi":"10.1109/IJCNN52387.2021.9534229","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9534229","url":null,"abstract":"A proper representation of predictive uncertainty is vital for deep neural networks (DNNs) to be applied in safety-critical domains such as medical diagnosis and self-driving. State-of-the-art (SOTA) variational inference approximation techniques provide a theoretical framework for modeling uncertainty, however, they have not been proven to work on large and deep networks with practical computation. In this study, we develop a layerwise approximation with a local reparameterization technique to efficiently perform sophisticated variational Bayesian inference on very deep SOTA convolutional neural networks (CNNs) (VGG16, ResNet variants, DenseNet). Theoretical analysis is presented to justify that the layerwise approach remains a Bayesian neural network. We further derive a SOTA $alpha$-divergence objective function to work with the layerwise approximate inference, addressing the concern of underestimating uncertainties by the Kullback-Leibler divergence. Empirical evaluation using MNIST, CIFAR-10, and CIFAR-100 datasets consistently shows that with our proposal, deep CNN models can have a better quality of predictive uncertainty than Monte Carlo-dropout in detecting in-domain misclassification and excel in out-of-distribution detection.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121240807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring the Effect of Dynamic Drive Balancing in Open-ended Learning Robots 探讨动态驱动平衡在开放式学习机器人中的作用
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9534137
A. Romero, F. Bellas, R. Duro
{"title":"Exploring the Effect of Dynamic Drive Balancing in Open-ended Learning Robots","authors":"A. Romero, F. Bellas, R. Duro","doi":"10.1109/IJCNN52387.2021.9534137","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9534137","url":null,"abstract":"This paper seeks to explore the effect and possibilities of autonomously balancing drives in a motivational architecture aimed at open-ended learning robots. These types of robots are very useful in unconstrained human robot interaction settings or when uncontrolled dynamic scenarios that are unknown at design time must be addressed. Designing a robot under these conditions implies that it must be endowed with some primary operational purpose and some additional self-preservation objectives whose fulfillment depend on the characteristics of the particular domain it is facing each moment in time. Domains that are not known beforehand and for which no a priori goal or skill structure can be designed in. Thus, an approach to the design and engineering of motivational structures to endow robots with specific purposes is proposed and tested here. We concentrate on the drive structure of a motivational system and the effects of its autonomous adaptation to changing circumstances. To provide for this adaptation, a simple evolutionary strategy is defined for the autonomous regulation of multiple drives seeking to optimize long-term operation. The proposal is tested on a Baxter robot performing an industrial task and the results confirm the potential of autonomous dynamic drive balancing as a tool in open-ended settings.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127465895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structure and Randomness in Planning and Reinforcement Learning 计划和强化学习中的结构和随机性
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9533317
K. Czechowski, Piotr Januszewski, Piotr Kozakowski, Łukasz Kuciński, Piotr Milos
{"title":"Structure and Randomness in Planning and Reinforcement Learning","authors":"K. Czechowski, Piotr Januszewski, Piotr Kozakowski, Łukasz Kuciński, Piotr Milos","doi":"10.1109/IJCNN52387.2021.9533317","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533317","url":null,"abstract":"Planning in large state spaces inevitably needs to balance the depth and breadth of the search. It has a crucial impact on the performance of a planner and most manage this interplay implicitly. We present a novel method Shoot Tree Search (STS), which makes it possible to control this trade-off more explicitly. Our algorithm can be understood as an interpolation between two celebrated search mechanisms: MCTS and random shooting. It also lets the user control the bias-variance trade-off, akin to TD(n), but in the tree search context. In experiments on challenging domains, we show that STS can get the best of both worlds consistently achieving higher scores.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124968012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Bag of Tricks for Chinese Named Entity Recognition 中文命名实体识别技巧包
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9533296
Yao Xiao, Jingbo Peng, Luoyi Fu, Haisong Zhang
{"title":"Bag of Tricks for Chinese Named Entity Recognition","authors":"Yao Xiao, Jingbo Peng, Luoyi Fu, Haisong Zhang","doi":"10.1109/IJCNN52387.2021.9533296","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533296","url":null,"abstract":"Named entity recognition (NER) is an important and challenging task in natural language processing. In this paper, we investigate thoroughly about the advances of Chinese NER in recent years. We explore the validity of a wide range of approaches in the literature of NLP that may benefit NER. We further employ the effective ones, such as data augmentation, adversarial learning, cross-sentence context and cost-sensitive learning to improve the performance of our BERT-based backbone model. Empirical results show that our model with this bag of tricks outperforms previous state-of-the-art on Weibo and achieves competitive performance on MSRA. Our code is publicly available11https://github.com/ccoay/bag-ner.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125045015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Agent Combat in Non-Stationary Environments 非静止环境下的多智能体战斗
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9534036
Shengang Li, Haoang Chi, Tao Xie
{"title":"Multi-Agent Combat in Non-Stationary Environments","authors":"Shengang Li, Haoang Chi, Tao Xie","doi":"10.1109/IJCNN52387.2021.9534036","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9534036","url":null,"abstract":"Multi-agent combat is a combat scenario in multiagent reinforcement learning (MARL). In this combat, agents use reinforcement learning methods to learn optimal policies. Actually, policy may be changed, which leads to a non-stationary environment. In this case, it is difficult to predict opponents' policies. Many reinforcement learning methods try to solve nonstationary problems. Most of the previous works put all agents into a frame and model their policies to deal with non-stationarity of environments. But, in a combat environment, opponents can not be in the same frame as our agents. We group opponents and our agents into two frames, only considering opponents as a part of the environment. In this paper, we focus on the problem of modelling opponents' policies in non-stationary environments. To solve this problem, we propose an algorithm called Additional Opponent Characteristics Multi-agent Deep Deterministic Policy Gradient (AOC-MADDPG) with the following contributions: (1) We propose a new actor-critic framework to deal with nonstationarity of environments in MARL, so that agents can adapt to more complex environments. (2) A model for opponents' policies is built by introducing observations and actions of the opponents into the critic network as additional characteristics. We evaluate our AOC-MADDPG algorithm in two multi-agent combat environments. As a result, our approach significantly outperforms the baseline. Agents trained by our method can get higher rewards in non-stationary environments.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"456 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125793357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Self-Supervised Learning Framework for Sequential Recommendation
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9534405
Renqi Jia, Xu Bai, Xiaofei Zhou, Shirui Pan
{"title":"A Self-Supervised Learning Framework for Sequential Recommendation","authors":"Renqi Jia, Xu Bai, Xiaofei Zhou, Shirui Pan","doi":"10.1109/IJCNN52387.2021.9534405","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9534405","url":null,"abstract":"Sequential recommendation that aims to predict user preference with historical user interactions becomes one of the most popular tasks in the recommendation area. The existing methods concentrated on user's sequential features among exposed items have achieved good performance. However, they only rely on single item prediction optimization to learn data representation, which ignores the association between context data and sequence data. In this paper, we propose a novel self-supervised learning based sequential recommendation network (SSLRN), which contrastively learns data correlation to promote data representation of users and items. We design two auxiliary contrastive learning tasks to regularize user and item representation based on mutual information maximization (MIM). In particular, the item contrastive learning captures sequential contrast feature with sequence-item MIM, and the user contrastive learning regularizes user latent representation with user-item MIM. We evaluate our model on five real-world datasets and the experimental results show that the proposed framework significantly and consistently outperforms state-of-the-art sequential recommendation techniques.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125842114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Prior Probability Estimation in Dynamically Imbalanced Data Streams 动态不平衡数据流中的先验概率估计
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9533795
Joanna Komorniczak, Paweł Zyblewski, Paweł Ksieniewicz
{"title":"Prior Probability Estimation in Dynamically Imbalanced Data Streams","authors":"Joanna Komorniczak, Paweł Zyblewski, Paweł Ksieniewicz","doi":"10.1109/IJCNN52387.2021.9533795","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533795","url":null,"abstract":"Despite the fact that real-life data streams may often be characterized by the dynamic changes in the prior class probabilities, there is a scarcity of articles trying to clearly describe and classify this problem as well as suggest new methods dedicated to resolving this issue. The following paper aims to fill this gap by proposing a novel data stream taxonomy defined in the context of prior class probability and by introducing the Dynamic Statistical Concept Analysis (DSCA) - prior probability estimation algorithm. The proposed method was evaluated using computer experiments carried out on 100 synthetically generated data streams with various class imbalance characteristics. The obtained results, supported by statistical analysis, confirmed the usefulness of the proposed solution, especially in the case of discrete dynamically imbalanced data streams (DDIS).","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125898512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Mastering the Game of Amazons Fast by Decoupling Network Learning 通过解耦网络学习快速掌握亚马逊游戏
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9534274
G. Q. Zhang, Xiaoyang Chen, Ruidong Chang, Yuhang Zhang, Cong Wang, Luyi Bai, Junwei Wang, Changming Xu
{"title":"Mastering the Game of Amazons Fast by Decoupling Network Learning","authors":"G. Q. Zhang, Xiaoyang Chen, Ruidong Chang, Yuhang Zhang, Cong Wang, Luyi Bai, Junwei Wang, Changming Xu","doi":"10.1109/IJCNN52387.2021.9534274","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9534274","url":null,"abstract":"In this work, we propose a deep reinforcement learning (DRL) algorithm DoubleJump which can master the game of Amazons efficiently. To address the bottleneck problem of sparse supervision signal in DRL, we split the neural network into rule network and skill network, using huge amounts of inexpensive data with game rule information and scarce data containing game skill information to train two networks respectively. Besides, we split the three sub-actions of each action into independent states during Monte-Carlo tree search (MCTS), to improve the probability of finding the global optimal state and reduce the average branching factor. The experimental results show our algorithm reaches about 130:70 in the zero-knowledge learning compared with the AlphaGo Zero algorithm, significantly improves the learning speed, and then alleviates the severe dependence on computing resources.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"163 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125913222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Graph-based Visual Manipulation Relationship Reasoning in Object-Stacking Scenes 对象堆叠场景中基于图的视觉操作关系推理
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9534389
Guoyu Zuo, Jiayuan Tong, Hongxing Liu, Wenbai Chen, Jianfeng Li
{"title":"Graph-based Visual Manipulation Relationship Reasoning in Object-Stacking Scenes","authors":"Guoyu Zuo, Jiayuan Tong, Hongxing Liu, Wenbai Chen, Jianfeng Li","doi":"10.1109/IJCNN52387.2021.9534389","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9534389","url":null,"abstract":"In object-stacking scenes, robotic manipulation is one of the most important research topics in robotics. It is particularly significant to reason object relationships and obtain intelligent manipulation order for more advanced interaction between the robot and the environment. However, many existing methods focus on individual object features and ignore contextual information, leading to great challenges in efficiently reasoning manipulation relationship. In this paper, we introduce a novel graph-based visual manipulation relationship reasoning architecture that directly outputs object relationships and manipulation order. Our model first extracts features and detects objects from RGB images, and then adopts Graph Convolutional Network (GCN) to collect contextual information between objects. Moreover, a relationship filtering network is built to reduce object pairs before reasoning and improve the efficiency of relation reasoning. The experiments on the Visual Manipulation Relationship Dataset (VMRD) show that our model significantly outperforms previous methods on reasoning object relationships in obiect-stackina scenes.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126076786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Strategic Weight Refinement Maneuver for Convolutional Neural Networks 卷积神经网络的一种策略权重细化策略
2021 International Joint Conference on Neural Networks (IJCNN) Pub Date : 2021-07-18 DOI: 10.1109/IJCNN52387.2021.9533359
Patrick Sharma, Adarsh Karan Sharma, Dinesh Kumar, Anuraganand Sharma
{"title":"A Strategic Weight Refinement Maneuver for Convolutional Neural Networks","authors":"Patrick Sharma, Adarsh Karan Sharma, Dinesh Kumar, Anuraganand Sharma","doi":"10.1109/IJCNN52387.2021.9533359","DOIUrl":"https://doi.org/10.1109/IJCNN52387.2021.9533359","url":null,"abstract":"Stochastic Gradient Descent algorithms (SGD) remain a popular optimizer for deep learning networks and have been increasingly used in applications involving large datasets producing promising results. SGD approximates the gradient on a small subset of training examples, randomly selected in every iteration during network training. This randomness leads to the selection of an inconsistent order of training examples resulting in ambiguous values to solve the cost function. This paper applies Guided Stochastic Gradient Descent (GSGD) - a variant of SGD in deep learning neural networks. GSGD minimizes the training loss and maximizes the classification accuracy by overcoming the inconsistent order of data examples in SGDs. It temporarily bypasses the inconsistent data instances during gradient computation and weight update, leading to better convergence at the rate of $O(frac{1}{rho T-})$. Previously, GSGD has only been used in the shallow learning networks like the logistic regression. We try to incorporate GSGD in deep learning neural networks like the Convolutional Neural Networks (CNNs) and evaluate the classification accuracy in comparison with the same networks trained with SGDs. We test our approach on benchmark image datasets. Our baseline results show GSGD leads to a better convergence rate and improves classification accuracy by up to 3% of standard CNNs.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126109700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信