Dănuţ-Vasile Giurgi , Mihreteab Negash Geletu , Thomas Josso-Laurain , Maxime Devanne , Jean-Philippe Lauffenburger , Jean Dezert
{"title":"基于原型证据神经网络的远程冲突管理","authors":"Dănuţ-Vasile Giurgi , Mihreteab Negash Geletu , Thomas Josso-Laurain , Maxime Devanne , Jean-Philippe Lauffenburger , Jean Dezert","doi":"10.1016/j.ijar.2025.109508","DOIUrl":null,"url":null,"abstract":"<div><div>Despite advances in integrating reasoning based on belief functions to generalise probabilistic representations, distance-to-prototype-based evidential deep neural networks are still emerging and require further consolidation. Existing studies in segmentation or classification tasks typically perform prior initialisation and do not address or mitigate the potential conflicts that may arise during fusion. This work investigates high-conflict scenarios within an evidential neural network for segmentation in autonomous driving, focusing on the distance-to-prototypes component, where prototypes, derived from feature maps, serve as sources of evidence and may yield contradictory information. Conflict is mitigated through parameter adjustments within the evidential reasoning, enhancing consistency before fusion. This enables more reliable data integration and a valid application of fusion rules and decision-making processes. The proposed rectification is validated on two prototype configurations of a deep evidential lidar-camera cross-fusion architecture, using two distance-based decision strategies and adapted metrics. The impact on the network's predictions is demonstrated through qualitative and quantitative results on road detection tasks with the KITTI dataset.</div></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"186 ","pages":"Article 109508"},"PeriodicalIF":3.0000,"publicationDate":"2025-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Conflict management in a distance to prototype-based evidential neural network\",\"authors\":\"Dănuţ-Vasile Giurgi , Mihreteab Negash Geletu , Thomas Josso-Laurain , Maxime Devanne , Jean-Philippe Lauffenburger , Jean Dezert\",\"doi\":\"10.1016/j.ijar.2025.109508\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Despite advances in integrating reasoning based on belief functions to generalise probabilistic representations, distance-to-prototype-based evidential deep neural networks are still emerging and require further consolidation. Existing studies in segmentation or classification tasks typically perform prior initialisation and do not address or mitigate the potential conflicts that may arise during fusion. This work investigates high-conflict scenarios within an evidential neural network for segmentation in autonomous driving, focusing on the distance-to-prototypes component, where prototypes, derived from feature maps, serve as sources of evidence and may yield contradictory information. Conflict is mitigated through parameter adjustments within the evidential reasoning, enhancing consistency before fusion. This enables more reliable data integration and a valid application of fusion rules and decision-making processes. The proposed rectification is validated on two prototype configurations of a deep evidential lidar-camera cross-fusion architecture, using two distance-based decision strategies and adapted metrics. The impact on the network's predictions is demonstrated through qualitative and quantitative results on road detection tasks with the KITTI dataset.</div></div>\",\"PeriodicalId\":13842,\"journal\":{\"name\":\"International Journal of Approximate Reasoning\",\"volume\":\"186 \",\"pages\":\"Article 109508\"},\"PeriodicalIF\":3.0000,\"publicationDate\":\"2025-06-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Approximate Reasoning\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0888613X25001495\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Approximate Reasoning","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0888613X25001495","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Conflict management in a distance to prototype-based evidential neural network
Despite advances in integrating reasoning based on belief functions to generalise probabilistic representations, distance-to-prototype-based evidential deep neural networks are still emerging and require further consolidation. Existing studies in segmentation or classification tasks typically perform prior initialisation and do not address or mitigate the potential conflicts that may arise during fusion. This work investigates high-conflict scenarios within an evidential neural network for segmentation in autonomous driving, focusing on the distance-to-prototypes component, where prototypes, derived from feature maps, serve as sources of evidence and may yield contradictory information. Conflict is mitigated through parameter adjustments within the evidential reasoning, enhancing consistency before fusion. This enables more reliable data integration and a valid application of fusion rules and decision-making processes. The proposed rectification is validated on two prototype configurations of a deep evidential lidar-camera cross-fusion architecture, using two distance-based decision strategies and adapted metrics. The impact on the network's predictions is demonstrated through qualitative and quantitative results on road detection tasks with the KITTI dataset.
期刊介绍:
The International Journal of Approximate Reasoning is intended to serve as a forum for the treatment of imprecision and uncertainty in Artificial and Computational Intelligence, covering both the foundations of uncertainty theories, and the design of intelligent systems for scientific and engineering applications. It publishes high-quality research papers describing theoretical developments or innovative applications, as well as review articles on topics of general interest.
Relevant topics include, but are not limited to, probabilistic reasoning and Bayesian networks, imprecise probabilities, random sets, belief functions (Dempster-Shafer theory), possibility theory, fuzzy sets, rough sets, decision theory, non-additive measures and integrals, qualitative reasoning about uncertainty, comparative probability orderings, game-theoretic probability, default reasoning, nonstandard logics, argumentation systems, inconsistency tolerant reasoning, elicitation techniques, philosophical foundations and psychological models of uncertain reasoning.
Domains of application for uncertain reasoning systems include risk analysis and assessment, information retrieval and database design, information fusion, machine learning, data and web mining, computer vision, image and signal processing, intelligent data analysis, statistics, multi-agent systems, etc.