{"title":"Theoretical aspects of robust SVM optimization in Banach spaces and Nash equilibrium interpretation","authors":"Mohammed Sbihi, Nicolas Couellan","doi":"10.1007/s10472-024-09931-z","DOIUrl":"10.1007/s10472-024-09931-z","url":null,"abstract":"<div><p>There are many real life applications where data can not be effectively represented in Hilbert spaces and/or where the data points are uncertain. In this context, we address the issue of binary classification in Banach spaces in presence of uncertainty. We show that a number of results from classical support vector machines theory can be appropriately generalized to their robust counterpart in Banach spaces. These include the representer theorem, strong duality for the associated optimization problem as well as their geometrical interpretation. Furthermore, we propose a game theoretical interpretation of the class separation problem when the underlying space is reflexive and smooth. The proposed Nash equilibrium formulation draws connections and emphasizes the interplay between class separation in machine learning and game theory in the general setting of Banach spaces.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1273 - 1293"},"PeriodicalIF":1.2,"publicationDate":"2024-02-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139767186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Learning preference representations based on Choquet integrals for multicriteria decision making","authors":"Margot Herin, Patrice Perny, Nataliya Sokolovska","doi":"10.1007/s10472-024-09930-0","DOIUrl":"10.1007/s10472-024-09930-0","url":null,"abstract":"<div><p>This paper concerns preference elicitation and learning of decision models in the context of multicriteria decision making. We propose an approach to learn a representation of preferences by a non-additive multiattribute utility function, namely a Choquet or bi-Choquet integral. This preference model is parameterized by one-dimensional utility functions measuring the attractiveness of consequences w.r.t. various point of views and one or two set functions (capacities) used to weight the coalitions and control the intensity of interactions among criteria, on the positive and possibly the negative sides of the utility scale. Our aim is to show how we can successively learn marginal utilities from properly chosen preference examples and then learn where the interactions matter in the overall model. We first present a preference elicitation method to learn spline representations of marginal utilities on every component of the model. Then we propose a sparse learning approach based on adaptive <span>(L_1)</span>-regularization for determining a compact Möbius representation fitted to the observed preferences. We present numerical tests to compare different regularization methods. We also show the advantages of our approach compared to basic methods that do not seek sparsity or that force sparsity a priori by requiring <i>k</i>-additivity.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 6","pages":"1511 - 1544"},"PeriodicalIF":1.2,"publicationDate":"2024-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139767178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Arwa Khannoussi, Alexandru-Liviu Olteanu, Patrick Meyer, Bastien Pasdeloup
{"title":"A metaheuristic for inferring a ranking model based on multiple reference profiles","authors":"Arwa Khannoussi, Alexandru-Liviu Olteanu, Patrick Meyer, Bastien Pasdeloup","doi":"10.1007/s10472-024-09926-w","DOIUrl":"10.1007/s10472-024-09926-w","url":null,"abstract":"<div><p>In the context of Multiple Criteria Decision Aiding, decision makers often face problems with multiple conflicting criteria that justify the use of preference models to help advancing towards a decision. In order to determine the parameters of these preference models, preference elicitation makes use of preference learning algorithms, usually taking as input holistic judgments, i.e., overall preferences on some of the alternatives, expressed by the decision maker. Tools to achieve this goal in the context of a ranking model based on multiple reference profiles are usually based on mixed-integer linear programming, Boolean satisfiability formulation or metaheuristics. However, they are usually unable to handle realistic problems involving many criteria and a large amount of input information. We propose here an evolutionary metaheuristic in order to address this issue. Extensive experiments illustrate its ability to handle problem instances that previous proposals cannot.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 6","pages":"1467 - 1488"},"PeriodicalIF":1.2,"publicationDate":"2024-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139767244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Alexandre Niveau, Hector Palacios, Sergej Scheck, Bruno Zanuttini
{"title":"A knowledge compilation perspective on queries and transformations for belief tracking","authors":"Alexandre Niveau, Hector Palacios, Sergej Scheck, Bruno Zanuttini","doi":"10.1007/s10472-023-09908-4","DOIUrl":"10.1007/s10472-023-09908-4","url":null,"abstract":"<div><p>Nondeterministic planning is the process of computing plans or policies of actions achieving given goals, when there is nondeterministic uncertainty about the initial state and/or the outcomes of actions. This process encompasses many precise computational problems, from classical planning, where there is no uncertainty, to contingent planning, where the agent has access to observations about the current state. Fundamental to these problems is belief tracking, that is, obtaining information about the current state after a history of actions and observations. At an abstract level, belief tracking can be seen as maintaining and querying the current belief state, that is, the set of states consistent with the history. We take a knowledge compilation perspective on these processes, by defining the queries and transformations which pertain to belief tracking. We study them for propositional domains, considering a number of representations for belief states, actions, observations, and goals. In particular, for belief states, we consider explicit propositional representations with and without auxiliary variables, as well as implicit representations by the history itself; and for actions, we consider propositional action theories as well as ground PDDL and conditional STRIPS. For all combinations, we investigate the complexity of relevant queries (for instance, whether an action is applicable at a belief state) and transformations (for instance, revising a belief state by an observation); we also discuss the relative succinctness of representations. Though many results show an expected tradeoff between succinctness and tractability, we identify some interesting combinations. We also discuss the choice of representations by existing planners in light of our study.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1113 - 1159"},"PeriodicalIF":1.2,"publicationDate":"2024-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139657708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Francine Chen, Yanxia Zhang, Minh Nguyen, Matt Klenk, Charlene Wu
{"title":"Personalized choice prediction with less user information","authors":"Francine Chen, Yanxia Zhang, Minh Nguyen, Matt Klenk, Charlene Wu","doi":"10.1007/s10472-024-09927-9","DOIUrl":"10.1007/s10472-024-09927-9","url":null,"abstract":"<div><p>While most models of human choice are linear to ease interpretation, it is not clear whether linear models are good models of human decision making. And while prior studies have investigated how task conditions and group characteristics, such as personality or socio-demographic background, influence human decisions, no prior works have investigated how to use less personal information for choice prediction. We propose a deep learning model based on self-attention and cross-attention to model human decision making which takes into account both subject-specific information and task conditions. We show that our model can consistently predict human decisions more accurately than linear models and other baseline models while remaining interpretable. In addition, although a larger amount of subject specific information will generally lead to more accurate choice prediction, collecting more surveys to gather subject background information is a burden to subjects, as well as costly and time-consuming. To address this, we introduce a training scheme that reduces the number of surveys that must be collected in order to achieve more accurate predictions.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 6","pages":"1489 - 1509"},"PeriodicalIF":1.2,"publicationDate":"2024-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10472-024-09927-9.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139648618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dmitry Semenov, Alexander Koldanov, Petr Koldanov, Panos Pardalos
{"title":"Clique detection with a given reliability","authors":"Dmitry Semenov, Alexander Koldanov, Petr Koldanov, Panos Pardalos","doi":"10.1007/s10472-024-09928-8","DOIUrl":"10.1007/s10472-024-09928-8","url":null,"abstract":"<div><p>In this paper we propose a new notion of a clique reliability. The clique reliability is understood as the ratio of the number of statistically significant links in a clique to the number of edges of the clique. This notion relies on a recently proposed original technique for separating inferences about pairwise connections between vertices of a network into significant and admissible ones. In this paper, we propose an extension of this technique to the problem of clique detection. We propose a method of step-by-step construction of a clique with a given reliability. The results of constructing cliques with a given reliability using data on the returns of stocks included in the Dow Jones index are presented.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"93 1","pages":"173 - 186"},"PeriodicalIF":1.2,"publicationDate":"2024-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139578602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fernando Díaz-del-Río, Helena Molina-Abril, Pedro Real, Darian Onchis, Sergio Blanco-Trejo
{"title":"Parallel homological calculus for 3D binary digital images","authors":"Fernando Díaz-del-Río, Helena Molina-Abril, Pedro Real, Darian Onchis, Sergio Blanco-Trejo","doi":"10.1007/s10472-023-09913-7","DOIUrl":"10.1007/s10472-023-09913-7","url":null,"abstract":"<div><p>Topological representations of binary digital images usually take into consideration different adjacency types between colors. Within the cubical-voxel 3D binary image context, we design an algorithm for computing the isotopic model of an image, called (<b>6</b>, <b>26</b>)-Homological Region Adjacency Tree ((<b>6</b>, <b>26</b>)-<i>Hom-Tree</i>). This algorithm is based on a flexible graph scaffolding at the inter-voxel level called Homological Spanning Forest model (HSF). <i>Hom-Trees</i> are edge-weighted trees in which each node is a maximally connected set of constant-value voxels, which is interpreted as a subtree of the HSF. This representation integrates and relates the homological information (connected components, tunnels and cavities) of the maximally connected regions of constant color using 6-adjacency and 26-adjacency for black and white voxels, respectively (the criteria most commonly used for 3D images). The Euler-Poincaré numbers (which may as well be computed by counting the number of cells of each dimension on a cubical complex) and the connected component labeling of the foreground and background of a given image can also be straightforwardly computed from its Hom-Trees. Being <span>(I_D)</span> a 3D binary well-composed image (where <i>D</i> is the set of black voxels), an almost fully parallel algorithm for constructing the <i>Hom-Tree</i> via HSF computation is implemented and tested here. If <span>(I_D)</span> has <span>(m_1{times } m_2{times } m_3)</span> voxels, the time complexity order of the reproducible algorithm is near <span>(O(log (m_1{+}m_2{+}m_3)))</span>, under the assumption that a processing element is available for each cubical voxel. Strategies for using the compressed information of the <i>Hom-Tree</i> representation to distinguish two topologically different images having the same homological information (Betti numbers) are discussed here. The topological discriminatory power of the <i>Hom-Tree</i> and the low time complexity order of the proposed implementation guarantee its usability within machine learning methods for the classification and comparison of natural 3<i>D</i> images.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 1","pages":"77 - 113"},"PeriodicalIF":1.2,"publicationDate":"2024-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10472-023-09913-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139578597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Weighted and Choquet (L^p) distance representation of comparative dissimilarity relations on fuzzy description profiles","authors":"Giulianella Coletti, Davide Petturiti, Bernadette Bouchon-Meunier","doi":"10.1007/s10472-024-09924-y","DOIUrl":"10.1007/s10472-024-09924-y","url":null,"abstract":"<div><p>We consider comparative dissimilarity relations on pairs on fuzzy description profiles, the latter providing a fuzzy set-based representation of pairs of objects. Such a relation expresses the idea of “no more dissimilar than” and is used by a decision maker when performing a case-based decision task under vague information. We first limit ourselves to those relations admitting a weighted <span>(varvec{L}^p)</span> distance representation, for which we provide an axiomatic characterization in case the relation is complete, transitive and defined on the entire space of pairs of fuzzy description profiles. Next, we switch to the more general class of comparative dissimilarity relations representable by a Choquet <span>(varvec{L}^p)</span> distance, parameterized by a completely alternating normalized capacity.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 6","pages":"1407 - 1436"},"PeriodicalIF":1.2,"publicationDate":"2024-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10472-024-09924-y.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139562232","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dimitrios I. Diochnos, Martin Charles Golumbic, Frederick Hoffman
{"title":"ISAIM-2022: international symposium on artificial intelligence and mathematics","authors":"Dimitrios I. Diochnos, Martin Charles Golumbic, Frederick Hoffman","doi":"10.1007/s10472-024-09922-0","DOIUrl":"10.1007/s10472-024-09922-0","url":null,"abstract":"","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 1","pages":"1 - 4"},"PeriodicalIF":1.2,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139611902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Stability of accuracy for the training of DNNs via the uniform doubling condition","authors":"Yitzchak Shmalo","doi":"10.1007/s10472-023-09919-1","DOIUrl":"10.1007/s10472-023-09919-1","url":null,"abstract":"<div><p>We study the stability of accuracy during the training of deep neural networks (DNNs). In this context, the training of a DNN is performed via the minimization of a cross-entropy loss function, and the performance metric is accuracy (the proportion of objects that are classified correctly). While training results in a decrease of loss, the accuracy does not necessarily increase during the process and may sometimes even decrease. The goal of achieving stability of accuracy is to ensure that if accuracy is high at some initial time, it remains high throughout training. A recent result by Berlyand, Jabin, and Safsten introduces a doubling condition on the training data, which ensures the stability of accuracy during training for DNNs using the absolute value activation function. For training data in <span>(mathbb {R}^n)</span>, this doubling condition is formulated using slabs in <span>(mathbb {R}^n)</span> and depends on the choice of the slabs. The goal of this paper is twofold. First, to make the doubling condition uniform, that is, independent of the choice of slabs. This leads to sufficient conditions for stability in terms of training data only. In other words, for a training set <i>T</i> that satisfies the uniform doubling condition, there exists a family of DNNs such that a DNN from this family with high accuracy on the training set at some training time <span>(t_0)</span> will have high accuracy for all time <span>(t>t_0)</span>. Moreover, establishing uniformity is necessary for the numerical implementation of the doubling condition. We demonstrate how to numerically implement a simplified version of this uniform doubling condition on a dataset and apply it to achieve stability of accuracy using a few model examples. The second goal is to extend the original stability results from the absolute value activation function to a broader class of piecewise linear activation functions with finitely many critical points, such as the popular Leaky ReLU.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 2","pages":"439 - 483"},"PeriodicalIF":1.2,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139506297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}