{"title":"Robust support vector machine based on the bounded asymmetric least squares loss function and its applications in noise corrupted data","authors":"Jiaqi Zhang, Hu Yang","doi":"10.1016/j.aei.2025.103371","DOIUrl":null,"url":null,"abstract":"<div><div>The support vector machine (SVM) is a popular machine learning tool that has achieved great success in various fields, but its performance is significantly disturbed on noise corrupted datasets. In this paper, motivated by the bounded quantile loss function, based on the relationship of the expectile and asymmetric least squares loss function, we propose the bounded asymmetric least squares loss function (<span><math><msub><mrow><mi>L</mi></mrow><mrow><mi>b</mi><mi>a</mi><mi>l</mi><mi>s</mi></mrow></msub></math></span> loss function). The <span><math><msub><mrow><mi>L</mi></mrow><mrow><mi>b</mi><mi>a</mi><mi>l</mi><mi>s</mi></mrow></msub></math></span> loss function is an extension of the asymmetric least squares loss function. <span><math><msub><mrow><mi>L</mi></mrow><mrow><mi>b</mi><mi>a</mi><mi>l</mi><mi>s</mi></mrow></msub></math></span> loss function inherits the good properties from the asymmetric least squares loss function, such as asymmetric and differentiable. Further, <span><math><msub><mrow><mi>L</mi></mrow><mrow><mi>b</mi><mi>a</mi><mi>l</mi><mi>s</mi></mrow></msub></math></span> loss function is more robust to noise in classification and regression problems. Next, we propose two models based on <span><math><msub><mrow><mi>L</mi></mrow><mrow><mi>b</mi><mi>a</mi><mi>l</mi><mi>s</mi></mrow></msub></math></span> loss function, namely, BALS-SVM and BALS-SVR. The <span><math><msub><mrow><mi>L</mi></mrow><mrow><mi>b</mi><mi>a</mi><mi>l</mi><mi>s</mi></mrow></msub></math></span> loss function is a non-convex loss function which makes it difficult to optimize. Thus, we design a clipping dual coordinate descent (clipDCD) based half-quadratic algorithm for solving the proposed models. We further find that BALS-SVM and BALS-SVR can be viewed as iterative weighted asymmetric least squares loss function based support vector machines and support vector regression, which enhances the interpretability of the models. Finally, we provide a theoretical analysis of the model based on a general framework of bounded loss function, mainly including Fisher consistency and noise insensitivity. Meanwhile, theoretical guarantees are provided for the proposed models. The results on the simulated dataset and the 14 classification and 11 regression benchmark dataset show that our method is superior compared to the classical methods and some state-of-the-art methods, especially on the noise corrupted dataset. The statistical tests further confirm this fact. Experiments on the Fashion MNIST dataset and gene expression dataset further illustrate that our proposed model also performs well in real environments.</div></div>","PeriodicalId":50941,"journal":{"name":"Advanced Engineering Informatics","volume":"65 ","pages":"Article 103371"},"PeriodicalIF":8.0000,"publicationDate":"2025-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advanced Engineering Informatics","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1474034625002642","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The support vector machine (SVM) is a popular machine learning tool that has achieved great success in various fields, but its performance is significantly disturbed on noise corrupted datasets. In this paper, motivated by the bounded quantile loss function, based on the relationship of the expectile and asymmetric least squares loss function, we propose the bounded asymmetric least squares loss function ( loss function). The loss function is an extension of the asymmetric least squares loss function. loss function inherits the good properties from the asymmetric least squares loss function, such as asymmetric and differentiable. Further, loss function is more robust to noise in classification and regression problems. Next, we propose two models based on loss function, namely, BALS-SVM and BALS-SVR. The loss function is a non-convex loss function which makes it difficult to optimize. Thus, we design a clipping dual coordinate descent (clipDCD) based half-quadratic algorithm for solving the proposed models. We further find that BALS-SVM and BALS-SVR can be viewed as iterative weighted asymmetric least squares loss function based support vector machines and support vector regression, which enhances the interpretability of the models. Finally, we provide a theoretical analysis of the model based on a general framework of bounded loss function, mainly including Fisher consistency and noise insensitivity. Meanwhile, theoretical guarantees are provided for the proposed models. The results on the simulated dataset and the 14 classification and 11 regression benchmark dataset show that our method is superior compared to the classical methods and some state-of-the-art methods, especially on the noise corrupted dataset. The statistical tests further confirm this fact. Experiments on the Fashion MNIST dataset and gene expression dataset further illustrate that our proposed model also performs well in real environments.
期刊介绍:
Advanced Engineering Informatics is an international Journal that solicits research papers with an emphasis on 'knowledge' and 'engineering applications'. The Journal seeks original papers that report progress in applying methods of engineering informatics. These papers should have engineering relevance and help provide a scientific base for more reliable, spontaneous, and creative engineering decision-making. Additionally, papers should demonstrate the science of supporting knowledge-intensive engineering tasks and validate the generality, power, and scalability of new methods through rigorous evaluation, preferably both qualitatively and quantitatively. Abstracting and indexing for Advanced Engineering Informatics include Science Citation Index Expanded, Scopus and INSPEC.